Citizen science is an important approach for ecological studies that require county-scale or broader geographic coverage. Here we report on project management, technology, and the variety of roles and levels of engagement of citizen scientists in the Soundscapes to Landscapes project, focused in Sonoma County, California, USA. The project uses low-cost autonomous recording units (ARUs) and bioacoustic analysis to monitor bird diversity and soundscape components at a regional scale. In five years, 259 citizen scientists collectively volunteered 8,390 hours on a range of tasks, with 40% field work and 41% bioacoustic reference data collection, but also including geographic information systems, social media, and data upload. Citizen scientists were a mix of expert collaborators, community volunteers, and undergraduate students. In five distinct field campaigns, 141 citizen scientists deployed recording devices during the breeding bird season at 1,281 sites on public and private lands, and collected 12,431 hours of raw audio recordings. For bioacoustic analysis, we used a custom web-based citizen science interface to produce labeled reference data, through which seven expert citizen scientists and 105 user-level citizen scientists produced 230,066 labeled audio clips. We found that word-of-mouth and personal connections were the best strategies for recruitment and retainment of citizen scientists. Levels of engagement among citizen scientists varied, and community volunteers who participated in more than one task contributed more time to the project overall. Undergraduate students were a valuable citizen scientist group in the project, contributing 48% of the total citizen scientist effort and were particularly important for field work success.
Citizen science is an important approach for collecting in situ biodiversity and ecological information such as species distributions, population abundance, phenology, and ecosystem productivity, with broader spatial and temporal sampling than is economically or logistically feasible with standard research methods (Bonney et al. 2009; Bonney et al. 2014; Dickinson et al. 2012). The growth of web-based informatics and location-enabled mobile technologies has helped simplify, coordinate, and manage surveys by citizen scientists, allowing collection of large amounts of spatially-located ecological data stored in centralized databases (Catlin-Groves 2012; Dickinson et al. 2012). For example, eBird is a well-established global bird monitoring project that uses web- and mobile-based data entry to archive millions of citizen scientist bird sightings per year (Sullivan et al. 2014). However, avian monitoring projects using citizen science, such as eBird or regionally coordinated breeding bird atlas projects, rely on contributors to accurately identify species by sight or vocalization, which limits participation and requires additional verification steps (Catlin-Groves 2012; Dickinson et al. 2010; Robertson et al. 2010). Further, reliance on volunteer birdwatchers can lead to spatial clusters (e.g., in public areas) (Tang et al. 2021; Robertson et al. 2010; Boakes et al. 2010; Tulloch et al. 2013) and temporal sampling bias (e.g., on weekends) (Boakes et al. 2010; Funk and Richardson 2002; Robertson et al. 2010; Szabo et al. 2007), thereby limiting scientific applications (Zhang 2020).
Through online citizen science engagement platforms such as Zooniverse, citizen scientists identify target species (e.g., Snapshot Wisconsin, https://dnr.wisconsin.gov/topic/research/projects/snapshot), space phenomena (e.g., Disk Detective, https://www.zooniverse.org/projects/ssilverberg/disk-detective), or other patterns of interest in large visual, auditory, or audiovisual datasets (Tang et al. 2021). Yet many projects that use crowdsourcing to label and classify data offer a limited level of involvement by providing one or two repetitive tasks, which may decrease long-term citizen scientist engagement and thus require thousands of volunteers to obtain the data necessary for analysis (Raddick et al. 2010; Garbarino and Mason 2016). Citizen science projects that are most successful at sustaining citizen scientist participation are often those in which citizen scientists are involved in project design (Shirk et al. 2012; Bonney et al. 2014), have opportunities for increased responsibility and recognition (Dickinson et al. 2012), and are personally invested in the outcome (e.g., local water quality monitoring) (Capdevila et al. 2020; Shirk et al. 2012).
Passive acoustic monitoring of the environment can provide information on overall ecosystem status and change (Krause and Farina 2016; Pijanowski et al. 2011) as well as on sound-producing wildlife, including birds, amphibians, insects and mammals (Balantic and Donovan 2020; Gibb et al. 2019). Bioacoustic analysis allows automatic detection of bird presence with greater sampling in time and space than with traditional bird observations (Campos-Cerqueira and Aide 2016; Furnas and Callas 2015), removes the influence of human presence on animal vocalization during sampling, and reduces individual observer bias.
Soundscapes to Landscapes (S2L) (Soundscapes to Landscapes 2022) is a distributed, citizen science-based acoustic monitoring project that uses cost-effective mobile- and web-based technologies, autonomous recording units (ARUs), and bioacoustic analysis to monitor bird diversity and broad soundscape components of anthrophony (e.g., cars, airplanes), geophony (e.g., wind, rain), and biophony (e.g., birds, insects, mammals) at a countywide scale. Project citizen scientists are community volunteers and undergraduate interns that participate at varying levels of responsibility depending on skill and interest. Here we discuss project management and the multiple tasks that citizen scientists performed, including deploying and retrieving ARUs, identifying bird vocalizations and other sounds, uploading data, generating field maps, and promoting the project. While the biodiversity results from this project are still in active development, our goal is to provide useful information for developing and managing passive acoustic monitoring projects that seek to engage citizen scientists in multiple roles. We report on successes and challenges with citizen scientist engagement, differences in productivity levels of undergraduate students versus community volunteers, and lessons learned that are broadly applicable to citizen science projects.
Sound data collection for S2L was focused in Sonoma County, California, USA. Spanning 4,118 km2, the county has a variety of habitat types, including forest, shrubland, grassland, agricultural, and urban (Figure 1). Sonoma hosts a broad network of birdwatchers, college students, and conservation-minded organizations and stakeholders. The county is largely (~90%) privately owned, which allowed property owners to participate as citizen scientists by providing access to their land and optionally by deploying ARUs.
The S2L project involved citizen scientists, staff, and a science team distributed across academic and non-profit institutions. It required a variety of strategies for project management and for citizen scientist recruitment and engagement (Figure 2). We thus sought technological solutions to provide efficient, low-cost, secure, and collaborative tools for implementing and managing project components and data streams (Figure 3). We provide technology website links in Supplemental Table 1.
Overall project management was led by a core team of science investigators and a paid full-time project coordinator. We used Slack for communication between the science team and highly involved citizen scientists, and we stored sound recordings, meeting notes, code, reference data, and geospatial data in Google Drive (Figure 3).
This project required coordination of hundreds of citizen scientists, properties, landowners, deployments, and sample sites. To organize and track each workflow component, we used Airtable, a cloud-based interface that combines the features of spreadsheets with those of a simplified database (Figures 2 and 3). Additional information on Airtable is provided in Supplemental Text 1. The project coordinator facilitated data collection and tracking using built-in Airtable features such as web-based forms (e.g., for volunteers to provide contact information) and calendar views (e.g., for scheduling deployments).
We created a website (Soundscapes to Landscapes 2022) as an information base. The site includes general project information, science outcomes (e.g., presentations, publications), monthly newsletters, social media links, and a volunteer portal. All citizen scientists reported their hours via a form in the portal, and these self-reported hours data are used in the results. More details about the project website are provided in Supplemental Text 4.
We used several approaches to recruit citizen scientists to participate in S2L. We divide citizen scientists into two categories: 1) community volunteers, including property owners, and 2) student interns who received academic credit. We initially recruited three community volunteers with extensive knowledge of bird vocalizations to participate as official collaborators. This ensured that we had the expertise needed to successfully accomplish the project, and these collaborators also assisted with citizen scientist recruitment.
To recruit community volunteers, we used word-of-mouth, presentations to local birding groups (e.g., local Audubon chapter), the project website, and social media. To recruit undergraduate students, we gave presentations in college classrooms, advertised on college career boards, and sent announcements to targeted email lists. All citizen scientists signed up on our website using a simple Airtable form, which allowed them to choose desired activities. For online bioacoustics activities, we also used volunteer recruitment websites SciStarter and VolunteerMatch, which had a global reach.
Staff provided training days for ARU deployments prior to each annual field campaign. We engaged local citizen scientists in bioacoustics activities by hosting a series of in-person training events in which we presented a project overview, provided dinner, and assisted citizen scientists with bird call validations (Figure 4). During the COVID-19 pandemic when in-person gatherings were not possible, we provided a series of short training videos on YouTube to enable remote participation in the project.
We used two types of ARUs to collect sound recordings at sample sites (Figure 3: Field-based). In 2017 and 2018, we used ten Android-based smartphones (US$300/unit) with attached microphones and waterproof cases, using the Arbimon Touch app developed by Sieve Analytics. In 2019, we transitioned to the AudioMoth (Hill et al. 2019), which has been used in bird, amphibian, insect, and mammal research applications (Barber-Meyer et al. 2020; LeBien et al. 2020; Zhong et al. 2020). The AudioMoth costs about US$85 per unit with batteries and SD memory card, is easy to program, and has a simple data upload from the SD card. We programmed ARUs to sample 1 of every 10 minutes, thus providing temporal sampling through day and night. We chose not to record continuously as this would require our project to archive and process large amounts of data, would incur higher data storage costs, and our project goal was to capture species at a site level, not every instance of vocalization. Further, recordings typically spanned 3 to 4 days, with the goal of capturing more spatial than temporal variation in a field season.
We developed a stratified random sampling design based on land use, distance-based variables, and forest structure and chemical properties (Supplemental Text 2). Using ArcGIS Pro geographic information system (GIS) software, trained citizen scientists created property-scale maps for citizen scientists to use in the field (Figure 3: Special tasks). Field teams used the free smartphone application Gaia GPS to navigate to site locations and the ArcGIS Survey123 application to collect auxiliary site data, including location coordinates, date and time of deployment, property information, and photographs of the survey site (Figure 3: Field-based). We chose Survey123 because it was easy to configure and update with a web-based form, was available on both Apple iOS and Android mobile devices, and allows citizen scientists to transfer the data to ArcGIS Online where a service automatically updates a vector point GIS layer.
From 2017 to 2021, the S2L project deployed ARUs annually in field campaigns spanning late March to early July, capturing most of the breeding season when birds vocalize for mating and for defending their territory. We used two approaches to deploy ARUs to sampling locations: volunteer deploy and mail deploy. For the volunteer approach, a team of at least two citizen scientists traveled to public and private properties to deploy ARUs. The mail approach involved sending the ARUs, instructions, and a prepaid return box to private property owners to deploy on their own properties.
The citizen scientists found sample clips of bird vocalizations from 8,083 hours of audio recording data we collected in the first four years of the project (Figure 3: Cloud-based/Remote). We used the web-based Arbimon platform, which allows users to sort, visualize, listen to, and identify bird vocalizations (Aide et al. 2013). We partnered with Sieve Analytics (the platform creator) to design a citizen science interface for collecting the representative bird vocalizations needed for developing automated detection approaches with deep learning.
In the Arbimon citizen science interface, a volunteer with bird vocalization knowledge (hereafter, “expert”) delineates a bounding box representing a distinct bird vocalization within a 1-minute spectrogram, that is, a time versus frequency graphic depiction of a recording (Figure 4). The system then uses this template region of interest (ROI) and a pattern matching algorithm (LeBien et al. 2020) to find similar ROIs above a specified correlation threshold in our sound recordings. This results in hundreds to thousands of potential matches.
The Arbimon citizen science interface allows user- (citizen scientists with minimal or no bird vocalization knowledge) and expert-level citizen scientists to validate if matched ROIs include the bird vocalization in the template ROI by assessing the spectrogram or by listening to the sound clip (Figure 4). The system records validations of the same ROI among multiple user-level citizen scientists and provides a consensus vote. We chose a threshold of three user-level citizen scientist votes for either present or absent to reach a consensus, at which point the ROI was removed from further review. The expert citizen scientist can provide sole votes on ROIs to quickly boost the number of present ROIs, and can also review consensus validations to reduce false positives. This approach allowed our citizen scientists to develop ROI data for 54 bird species (Figure 3: Outputs).
Our goal was to collect ROI data to train a deep learning algorithm that would detect bird vocalizations and could be applied to recordings at the second level. To gain a realistic sense of model accuracy, we sought a per-second assessment of bird species within a sample of recordings, independent of ROIs. We selected a total of 905 1-minute recordings using a stratified random design (see Supplemental Text 3). Expert citizen scientists used Arbimon to visualize spectrograms and listen to each recording, and noted the start times of bird vocalizations in an Airtable base, creating a soundscape dataset (Figure 3: Outputs).
The project also sought to detect broad soundscape components of anthrophony, biophony, and geophony (ABG) in our recordings. User-level citizen scientists manually collected ROI data for ABG using two tools: Arbimon and Raven Lite from Cornell Lab of Ornithology. The Raven Lite software was not web-based, but was better suited for delineating ROIs around ABG than Arbimon, which was designed for biodiversity applications. Additionally, user-level citizen scientists noted start and stop times of ABG in a subset of 682 1-minute recordings from the soundscape dataset.
We used several approaches to report on project progress to help keep citizen scientists motivated and engaged (Figure 3: Reporting). First, we showed citizen scientist hours in a leaderboard on our project website (see Supplemental Text 4; Supplemental Figure 1). Further, we used Mailchimp to send a monthly email newsletter that included project updates and opportunities for citizen scientist participation, a featured bird-of-the-month with an example bird vocalization, a highlight of a citizen scientist, and a spotlight on a partner organization. We used the website to highlight publications or media content about our project, and we used Buffer to post regularly on social media with project updates, images, and announcements. Throughout the project, two citizen scientists (one community volunteer and one student intern) helped manage our social media accounts (Figure 3: Special tasks).
A total of 259 citizen scientists collectively contributed 8,390 hours over the duration of the study, primarily on 40% field work and 41% bioacoustic data collection. Student interns and citizen scientist collaborators were a driving force in accomplishing much of the citizen science tasks and contributed to 68% of total citizen scientist hours. Community volunteers contributed more time to the project if they participated in multiple tasks (Figure 5). Conversely, student contributions did not vary based on the number of tasks that they participated in (Figure 5). Student interns contributed more time to data upload, field work, and ABG ROI collection, while community volunteers contributed more time to all other tasks (Figure 6). The greatest number of citizen scientists participated in cloud-based/remote activities, resulting in the greatest amount of volunteer time dedicated to these activities.
In total, 141 citizen scientists contributed 3,687 hours to field work activities. Over five field campaigns, citizen scientists used 125 ARUs to collect sound recordings from 1,281 sites (Table 1). With volunteer deployments, ARU turnaround was 4 to 5 days, whereas mail deployment ARU turnaround was 2 to 3 weeks. During the study, we visited 192 properties (35% public versus 65% private land).
|VOLUNTEER DEPLOY||MAIL DEPLOY||ALL|
|Number of sites surveyed||1,130||151||1,281|
|Sites per volunteer hour||0.33||0.55||0.34|
|Sites per deployment (avg)||5.8||1.6||4.4|
|Hours of recordings||10,353||2,078||12,431|
|Hours of recordings per site (avg)||9.2||13.8||9.7|
|Number of days deployed per site (avg)||3.8||5.8||4|
|Number of deployment events||195||93||288|
Citizen scientists volunteered 3,448 hours on cloud-based/remote activities over the four years that the project was focused on collecting bioacoustic reference data. Seven expert citizen scientists with knowledge of bird vocalizations were critical to this phase and volunteered 1,715 hours to work on the following tasks: initial experimentation and testing of bird call modeling, finding representative bird vocalizations for pattern matching and reviewing ROI validations, and creating the soundscape dataset (Supplemental Text 3).
User-level citizen scientists volunteered 1,733 hours to: identify ABG soundscape components, validate ROIs in Arbimon, and manually collect a total of 5,396 ABG ROIs in Raven Lite. Hours came primarily from local citizen scientists, with ~9% from SciStarter and VolunteerMatch. Of 105 user-level citizen scientists, the top numbers of citizen scientists and validations were from citizen scientists recruited through colleges, word-of-mouth, and VolunteerMatch.
Our work with the Arbimon citizen science interface spanned 16 months and included 686,858 individual votes on pattern match results (average 400 votes per citizen scientist hour). There were 230,066 final bird vocalization ROIs validated by citizen scientists through consensus-based (76%) or expert (24%) methods (Figure 7; Supplemental Table 2). These ROIs included 54 species with a collective 51,906 clips indicating a species presence, ranging from 496 to 3,049 ROIs per species (average = 961). Since acquiring enough consensus-based present ROIs for a given bird species could be slow, citizen scientist experts helped move all species to completion faster; including 20 exclusively expert-validated species (Supplemental Table 2). In the end, 75% of our ROIs with species present were validated by experts (Figure 7).
The S2L project was successful at harnessing low-cost technology and the energy of citizen scientists interested in biodiversity conservation to acquire a large volume of acoustic recordings at a regional scale. Further, citizen scientists participated directly in the scientific process by collecting bird vocalization and soundscape component reference data for training and testing of deep learning algorithms for subsequent scientific analyses.
Of all outreach methods used to engage citizen scientists, word-of-mouth and local connections were the most successful in recruiting and maintaining participation. Targeted emails to local colleges attracted a large pool of students desiring field experience, and who ultimately completed the majority of volunteer deployments. Presenting to community groups was also effective in recruiting citizen scientists, particularly for bioacoustics-based activities. We had less success with online platforms VolunteerMatch, SciStarter, and social media, and citizen scientists recruited through these methods tended to be one-time participants.
Community volunteers were more likely to stay engaged in the project if they participated in more than one task. For example, several volunteers who started by doing field work also became interested in working with the collected audio data as they learned more about the project. Additionally, volunteers could utilize existing skills to participate in activities they were most interested in, like social media, bird identification, GIS, and programming. Because of this variety, many volunteers stayed engaged with the project even after their participation in a specific task declined. By engaging citizen scientists in multiple components of the project, it may help them develop a deeper sense of commitment by feeling co-ownership of the project and its outcomes (Pandya 2012).
It worked well to engage citizen scientists in the bioacoustics work using remote participation and in-person bird blitzes. Bird blitzes gave participants the advantage of meeting the project team, learning more about the science, and obtaining one-on-one training. Conversely, using a web-based interface enabled broader geographic citizen scientist recruitment as well as remote participation by locals who were unable to attend in-person events. Remote participation became a significant advantage during the COVID-19 pandemic and associated stay-at-home restrictions.
Roughly half of all citizen scientist hours in the project were from undergraduate students who received academic credit through internships with collaborator Point Blue Conservation Science. Engaging this citizen scientist group annually provided a cohort of students desiring field experience that could be expected to participate for the duration of the semester. Many students participated in only one annual field campaign, requiring us to provide training before each campaign. Notably, community volunteers ultimately required roughly the same amount of training hours as students (Figure 6: Meetings and trainings). The number of tasks that students participated in were not correlated with an increase in hours (Figure 5) because we assigned a range of tasks to give students a robust educational experience.
Besides these utilitarian benefits, a growing body of research links undergraduate involvement in citizen science projects and improved student learning outcomes, such as a better understanding of the scientific process, better connection to the natural world, and data literacy (see Hitchcock et al. 2021 in the special collection “Citizen Science in Higher Education” in Citizen Science: Theory and Practice). In S2L, we were able to engage students in multiple tasks, leveraging their existing knowledge in topics of social media, bird identification, and GIS. We did not set out to measure student learning outcomes, and thus have no quantitative data to assess how students were impacted by their experience—a limitation of many citizen science projects (NASEM 2018). However, ancillary evidence from students featured in our newsletter spotlight indicates that they were particularly motivated by the ability to explore the natural world, and they gained a sense of scientific discovery.
We sought a range of citizen scientist engagement in deploying ARUs with a driving goal to sample as much private land in the county as possible. We succeeded at reaching this goal; two thirds of sampled sites were located on private lands.
Although we envisioned mail deployment as a means to access more private lands than may be otherwise possible, we found that volunteer deployments were more efficient and provided the majority of our sound recordings. Although volunteer deployments required extensive coordination, citizen scientists were trained prior to conducting field work and generally needed little technical support once in the field. Further, the rapid ARU turnaround with volunteer deployments enabled us to redeploy the same devices to other sites quickly, whereas slow ARU turnaround with mail deployments limited redeployment. Although we provided detailed written instructions, as well as training videos in the fifth year, some property owners still found using the ARUs and smartphone applications challenging; they needed technical coaching or gave up entirely. Other citizen science projects have also found the technical design of applications and web-based interfaces to be a barrier for citizen scientist participation, and suggest that applications should be easy to learn and use (Benyei et al. 2020; Hobbs and White 2012; Martin et al. 2016).
We found AudioMoths to be low-cost and easy to program and deploy, and they replaced smartphones as the preferred ARUs in our project. One disadvantage was that AudioMoths lost their programming when the batteries were dislodged. It also may have been more efficient if there was an integrated smartphone application that could bundle functions of navigating to the sample site (e.g., Gaia GPS), logging location and auxiliary information (e.g., Survey 123), connecting to the AudioMoth via bluetooth to turn it on/off (e.g., currently a manual switch), and uploading recordings to our data repository via WiFi.
We used a custom citizen science interface in Arbimon for collecting our bird vocalization reference data. This is in contrast to many citizen science projects that use the Zooniverse platform for crowd-sourced, consensus-based voting, particularly within the domain of image interpretation (Cox et al. 2015; Willi et al. 2019). Arbimon allowed us to leverage web-based soundscape infrastructure and domain expertise at Sieve Analytics, and allowed integration with the system’s cloud-based pattern matching algorithm. Arbimon tools were easy for expert citizen scientists to use to identify representative ROI templates and run pattern matching jobs in a web-based environment that had access to all project recordings. The citizen science interface had an advantage in showing both site and temporal information for multiple matched ROIs on one page, making review more accurate and efficient.
Despite these benefits, Arbimon’s pattern matching algorithm results revealed that it can be inefficient for many species, producing many matches where the bird is absent, which ultimately wastes citizen scientist time. Therefore, we relied on experts to validate enough ROIs with birds present in order to proceed with deep learning development, which will be presented in a forthcoming paper.
As a start-up, science-based project relying on the academic and non-profit sector, we faced obstacles due to limited time and funding as well as a distributed work environment. We leveraged web-based tools (e.g., Slack, Google Drive, Airtable) for effective, cloud-based, low-cost collaboration and data management. However, owing to the significant time needed to develop our data processing pipeline, we did not reach the final outcomes at the time of completing field campaigns (Figure 2: Steps 7–10). One outcome that is now complete are landowner reports, which include the overall species richness and the number of detections of species and soundscape components through a 24-hr period for each site on a property. Additional project outcomes are regional maps of species distributions and ABG soundscape components using geospatial modeling. These deliverables are in progress and will be described in subsequent papers. The lag time in providing deliverables to citizen scientists and property owners potentially decreased long-term project enthusiasm, as consistent and timely feedback with citizen scientists often increases motivation for longer-term engagement, particularly when that feedback relates individual citizen scientist contributions to scientific outcomes (de Vries et al. 2019).
The development of low-cost sound ARUs allows passive acoustic sampling over more locations and time periods than previously available with more expensive equipment, and opens exciting opportunities to enlist citizen scientists in environmental monitoring. Further, web-based tools provide a platform for citizen scientists to assist with bioacoustic analysis at scale. In five years, the S2L project engaged 259 citizen scientists to collect 12,431 hours of audio recordings and validate 230,066 samples of bird vocalizations needed for automated species detection algorithms. Although S2L was focused on broad spatial sampling at the county scale, our approach of engaging undergraduate students and community volunteers in citizen science could be scaled to include sampling over larger regions and longer time periods. Offering a variety of tasks for citizen scientists to participate in was a successful engagement approach. However, there is no single platform that will meet all the data management, analysis, and social engagement needs of a citizen science bioacoustics project like S2L. Acoustic surveys are a cost-effective means to monitor biodiversity, thus the development of bioacoustic citizen science platforms is needed.
The sound recordings and reference data collected as part of Soundscapes to Landscapes have not yet been published. We intend to publish these data in a public depository upon completion of the project.
The supplementary file for this article can be found as follows:Supplemental Materials
We thank our citizen scientists for their time and dedication to making this project possible and for their devotion to bird conservation. In particular, we thank birder Bob Hasenick for his investment of time in bioacoustic analysis, volunteering more than 100 hours to the project, and Dylan Chapple and Phillip Carlos for initial S2L project management help. We also thank Sieve Analytics for implementing the Arbimon citizen science interface used in this project.
The Soundscapes to Landscapes project was funded by NASA’s Citizen Science for Earth Systems Program (CSESP) under cooperative agreement 80NSSC18M0107.
The authors have no competing interests to declare.
LS and MC designed and obtained funding for S2L. LS, MC, and RS coordinated project management. LS, MC, and RS performed data processing and analysis. DL, MT, TE, TT, TS, and WS were citizen scientist bird experts in bioacoustics work, and each volunteered more than 100 hours to the project. WS helped coordinate and train citizen scientists in bioacoustics work in years one and two. TS helped with field and citizen scientist coordination during year four in a temporary paid position. KC, a citizen scientist, worked on social media and bird call validations, and volunteered more than 100 hours to the project. LS, MC, and RS wrote the first draft of the paper. All co-authors commented on and approved the final manuscript.
Aide, TM, Corrada-Bravo, C, Campos-Cerqueira, M, Milan, C, Vega, G and Alvarez, R. 2013. Real-time bioacoustics monitoring and automated species identification. PeerJ, 1: e103. DOI: https://doi.org/10.7717/peerj.103
Balantic, CM and Donovan, TM. 2020. Statistical learning mitigation of false positives from template-detected data in automated acoustic wildlife monitoring. Bioacoustics, 29(3): 296–321. DOI: https://doi.org/10.1080/09524622.2019.1605309
Barber-Meyer, SM, Palacios, V, Marti-Domken, B and Schmidt, LJ. 2020. Testing a new passive acoustic recording unit to monitor wolves. Wildlife Society Bulletin, 44(3): 590–598. DOI: https://doi.org/10.1002/wsb.1117
Benyei, P, Pardo-de-Santayana, M, Aceituno-Mata, L, Calvet-Mir, L. Carrascosa-García, M, Rivera-Ferre, M, Perdomo-Molina, A and Reyes-García, V. 2020. Participation in citizen science: insights from the CONECT-e case study. Science, Technology, & Human Values: 0162243920948110. DOI: https://doi.org/10.1177/0162243920948110
Boakes, EH, McGowan, PJ, Fuller, RA, Chang-qing, D, Clark, NE, O’Connor, K and Mace, GM. 2010. Distorted views of biodiversity: spatial and temporal bias in species occurrence data. PLoS biology, 8(6): e1000385. DOI: https://doi.org/10.1371/journal.pbio.1000385
Bonney, R, Cooper, CB, Dickinson, J, Kelling, S, Phillips, T, Rosenberg, KV and Shirk, J. 2009. Citizen science: a developing tool for expanding science knowledge and scientific literacy. BioScience, 59(11): 977–984. DOI: https://doi.org/10.1525/bio.2009.59.11.9
Bonney, R, Shirk, JL, Phillips, TB, Wiggins, A, Ballard, HL, Miller-Rushing, AJ and Parrish, JK. 2014. Next steps for citizen science. Science, 343(6178): 1436–1437. DOI: https://doi.org/10.1126/science.1251554
Campos-Cerqueira, M and Aide, TM. 2016. Improving distribution data of threatened species by combining acoustic monitoring and occupancy modelling. Methods in Ecology and Evolution, 7(11): 1340–1348. DOI: https://doi.org/10.1111/2041-210X.12599
Capdevila, ASL, Kokimova, A, Ray, SS, Avellán, T, Kim, J and Kirschke, S. 2020. Success factors for citizen science projects in water quality monitoring. Science of the Total Environment, 728: 137843. DOI: https://doi.org/10.1016/j.scitotenv.2020.137843
Catlin-Groves, CL. 2012. The citizen science landscape: from volunteers to citizen sensors and beyond. International Journal of Zoology, 2012. DOI: https://doi.org/10.1155/2012/349630
Cox, J, Oh, EY, Simmons, B, Lintott, C, Masters, K, Greenhill, A, Graham, G and Holmes, K. 2015. Defining and measuring success in online citizen science: a case study of Zooniverse projects. Computing in Science & Engineering, 17(4): 28–41. DOI: https://doi.org/10.1109/MCSE.2015.65
de Vries, M, Land-Zandstra, A and Smeets, I. 2019. Citizen scientists’ preferences for communication of scientific output: a literature review. Citizen Science: Theory and Practice, 4(1): 2. DOI: https://doi.org/10.5334/cstp.136
Dickinson, JL, Shirk, J, Bonter, D, Bonney, R, Crain, RL, Martin, J, Phillips, T and Purcell, K. 2012. The current state of citizen science as a tool for ecological research and public engagement. Frontiers in Ecology and the Environment, 10(6): 291–297. DOI: https://doi.org/10.1890/110236
Dickinson, JL, Zuckerberg, B and Bonter, DN. 2010. Citizen science as an ecological research tool: challenges and benefits. Annual review of ecology, evolution, and systematics, 41: 149–172. DOI: https://doi.org/10.1146/annurev-ecolsys-102209-144636
Funk, VA and Richardson, KS. 2002. Systematic data in biodiversity studies: use it or lose it. Systematic biology, 51(2): 303–316. DOI: https://doi.org/10.1080/10635150252899789
Furnas, BJ and Callas, RL. 2015. Using automated recorders and occupancy models to monitor common forest birds across a large geographic region. The Journal of Wildlife Management, 79(2): 325–337. DOI: https://doi.org/10.1002/jwmg.821
Garbarino, J and Mason, CE. 2016. The power of engaging citizen scientists for scientific progress. Journal of microbiology & biology education, 17(1): 7–12. DOI: https://doi.org/10.1128/jmbe.v17i1.1052
Gibb, R, Browning, E, Glover-Kapfer, P and Jones, KE. 2019. Emerging opportunities and challenges for passive acoustics in ecological assessment and monitoring. Methods in Ecology and Evolution, 10(2): 169–185. DOI: https://doi.org/10.1111/2041-210X.13101
Hill, AP, Prince, P, Snaddon, JL, Doncaster, CP and Rogers, A. 2019. AudioMoth: A low-cost acoustic device for monitoring biodiversity and the environment. HardwareX, 6: e00073. DOI: https://doi.org/10.1016/j.ohx.2019.e00073
Hitchcock, C, Vance-Chalcraft, H and Aristeidou, M. 2021. Citizen science in higher education. Citizen Science: Theory and Practice, 6(1): 22. DOI: https://doi.org/10.5334/cstp.467
Hobbs, SJ and White, PC. 2012. Motivations and barriers in relation to community participation in biodiversity recording. Journal for Nature Conservation, 20(6): 364–373. DOI: https://doi.org/10.1016/j.jnc.2012.08.002
Krause, B and Farina, A. 2016. Using ecoacoustic methods to survey the impacts of climate change on biodiversity. Biological conservation, 195: 245–254. DOI: https://doi.org/10.1016/j.biocon.2016.01.013
LeBien, J, Zhong, M, Campos-Cerqueira, M, Velev, JP, Dodhia, R, Ferres, JL and Aide, TM. 2020. A pipeline for identification of bird and frog species in tropical soundscape recordings using a convolutional neural network. Ecological Informatics, 59: 101113. DOI: https://doi.org/10.1016/j.ecoinf.2020.101113
Martin, VY, Christidis, L, Lloyd, DJ and Pecl, G. 2016. Understanding drivers, barriers and information sources for public participation in marine citizen science. Journal of Science Communication, 15(2): A02. DOI: https://doi.org/10.22323/2.15020202
National Academies of Sciences, Engineering, and Medicine. 2018. Learning through citizen science: enhancing opportunities by design. Washington, DC: The National Academies Press. DOI: https://doi.org/10.17226/25183
Pandya, RE. 2012. A framework for engaging diverse communities in citizen science in the US. Frontiers in Ecology and the Environment, 10(6): 314–317. DOI: https://doi.org/10.1890/120007
Pijanowski, BC, Villanueva-Rivera, LJ, Dumyahn, SL, Farina, A, Krause, BL, Napoletano, BM, Gage, SH and Pieretti, N. 2011. Soundscape ecology: the science of sound in the landscape. BioScience, 61(3): 203–216. DOI: https://doi.org/10.1525/bio.2011.61.3.6
Raddick, MJ, Bracey, G, Gay, PL, Lintott, CJ, Murray, P, Schawinski, K, Szalay, AS and Vandenberg, J. 2010. Galaxy Zoo: exploring the motivations of citizen science volunteers. Astronomy Education Review, 9(1): n1. DOI: https://doi.org/10.3847/AER2009036
Robertson, MP, Cumming, GS and Erasmus, BFN. 2010. Getting the most out of atlas data. Diversity and Distributions, 16(3): 363–375. DOI: https://doi.org/10.1111/j.1472-4642.2010.00639.x
Shirk, JL, Ballard, HL, Wilderman, CC, Phillips, T, Wiggins, A, Jordan, R, McCallie, E, Minarchek, M, Lewenstein, BV, Krasny, ME and Bonney, R. 2012. Public participation in scientific research: a framework for deliberate design. Ecology and society, 17(2). DOI: https://doi.org/10.5751/ES-04705-170229
Soundscapes to Landscapes. 2022. Available at https://soundscapes2landscapes.org (Last accessed 31 January 2022).
Sullivan, BL, Aycrigg, JL, Barry, JH, Bonney, RE, Bruns, N, Cooper, CB, Damoulas, T, Dhondt, AA, Dietterich, T, Farnsworth, A and Fink, D. 2014. The eBird enterprise: an integrated approach to development and application of citizen science. Biological Conservation, 169: 31–40. DOI: https://doi.org/10.1016/j.biocon.2013.11.003
Szabo, JK, Davy, PJ, Hooper, MJ and Astheimer, LB. 2007. Predicting spatio-temporal distribution for eastern Australian birds using Birds Australia’s Atlas data: survey method, habitat and seasonal effects. Emu-Austral Ornithology, 107(2): 89–99. DOI: https://doi.org/10.1071/MU06020
Tang, B, Clark, JS and Gelfand, AE. 2021. Modeling spatially biased citizen science effort through the eBird database. Environmental and Ecological Statistics, 1–22. DOI: https://doi.org/10.1007/s10651-021-00508-1
Tulloch, AI, Mustin, K, Possingham, HP, Szabo, JK and Wilson, KA. 2013. To boldly go where no volunteer has gone before: predicting volunteer activity to prioritize surveys at the landscape scale. Diversity and Distributions, 19(4): 465–480. DOI: https://doi.org/10.1111/j.1472-4642.2012.00947.x
Willi, M, Pitman, RT, Cardoso, AW, Locke, C, Swanson, A, Boyer, A, Veldthuis, M and Fortson, L. 2019. Identifying animal species in camera trap images using deep learning and citizen science. Methods in Ecology and Evolution, 10(1): 80–91. DOI: https://doi.org/10.1111/2041-210X.13099
Zhang, G. 2020. Spatial and temporal patterns in volunteer data contribution activities: a case study of eBird. ISPRS International Journal of Geo-Information, 9(10): 597. DOI: https://doi.org/10.3390/ijgi9100597
Zhong, M, LeBien, J, Campos-Cerqueira, M, Dodhia, R, Ferres, JL, Velev, JP and Aide, TM. 2020. Multispecies bioacoustic classification using transfer learning of deep convolutional neural networks with pseudo-labeling. Applied Acoustics, 166: 107375. DOI: https://doi.org/10.1016/j.apacoust.2020.107375