SciCrowd: Towards a Hybrid, Crowd-Computing System for Supporting Research Groups in Academic Settings
- 3 Citations
- 497 Downloads
Abstract
The increasing amount of scholarly literature and the diversity of dissemination channels are challenging several fields and research communities. A continuous interplay between researchers and citizen scientists creates a vast set of possibilities to integrate hybrid, crowd-machine interaction features into crowd science projects for improving knowledge acquisition from large volumes of scientific data. This paper presents SciCrowd, an experimental crowd-powered system under development “from the ground up” to support data-driven research. The system combines automatic data indexing and crowd-based processing of data for detecting topic evolution by fostering a knowledge base of concepts, methods, and results categorized according to the particular needs of each field. We describe the prototype and discuss its main implications as a mixed-initiative approach for leveraging the analysis of academic literature.
Keywords
AI Crowd-computing hybrids Human computation Hybrid machine-crowd interaction Knowledge discovery Mixed-initiative systems Crowdsourcing1 Introduction
A major research path in today’s computing landscape has been focused on analyzing datasets whose volume, velocity, and variety are so extreme that the current automatic tools are inadequate for their accurate collection, management, and analysis in a reasonable amount of time [17]. As the literature increases at an exponential rate, scholars need to adapt their institutions and practices by improving searching, analytical, and filtering mechanisms to select the most relevant data sources. Nonetheless, manual labor applied on examining variances, correlating evidence, and compiling descriptive statistics is an exhaustive and subjective process [13]. At the forefront are the rapidly advancing techniques of AI such as machine learning running on large collections of data [1]. However, computer algorithms involve expensive training being prone to errors. Dunne and colleagues [14] go even further by arguing that the “current theories and tools are directed at finding a paper or website, not gaining an understanding of the key papers, authors, controversies, and hypotheses.” In this sense, research scholars interested in answering a particular question need intelligent systems robust enough to assist them in exploring always-evolving knowledge representations as a “science of science measurement” [19].
In this work, SciCrowd is introduced as a crowd-powered system prototype under development “from the ground up” towards an open participation model in which a large amount of researchers and crowd workers/volunteers can contribute for improving the way as research data are processed. We assume that crowd-computing hybrids constitute a valuable instrument to address some of the limitations of current applications by enabling us to identify new relations between research topics, authors, and groups [10]. Meaningful categories, filters, and summaries can be generated through mixed-initiative approaches combining human and machine intelligence [13]. Such synergies between mechanical and cognitive operators (the “crowd”) can contribute to create open knowledge bases supported by collaborative participation.
The paper is organized as follows. Section 2 provides an overview of key issues in the form of theoretical bridge between crowd and machine intelligence for enhancing research pursuits. Section 3 presents a brief conceptualization of the requirements for SciCrowd as an experimental human-centered hybrid system for scientific discovery. The implementation details are also discussed with emphasis on the SciCrowd’s prototype and architecture. This section also provides some exploratory remarks on the results achieved from a survey with crowdsourcing researchers. The paper finishes discussing limitations and future directions that need consideration in the deployment of a crowd-powered, self-organizing system for scientific discovery.
2 Related Work
Scientific knowledge production is being increasingly dependent on collaborative efforts and new forms of collaboration and data analysis are dawning as result of social-technical advances achieved in e-Science through CSCW research [2]. Various approaches have been proposed to leverage crowd work for improving the quality of data. Human crowds can combine Human Intelligence Tasks (HITs) and large-scale database systems such as MTurk1 to cope with the ambiguity, vagueness, heterogeneity, scale, timeliness, and complexity of scientific data [10]. Empirical studies have emphasized crowdsourcing efforts in domains such as taxonomy development [22], machine learning experiments using MTurk [20], and scientific discovery games (e.g., Foldit2) [6]. In addition, several crowd-powered systems have been presented in the literature. For instance, GATE Crowdsourcing Plugin [12] offers infrastructural support for supporting a massive pool of users on mapping corpora to microtasks while automatically generating crowdsourcing interfaces for classification and annotation. CrowdScape [21] measures crowd behavior and outputs though interactive visualization and mixed-initiative machine learning. Moreover, Apolo [8] allows interactive large graph sensemaking by combining machine learning and visualization. On the other hand, PANDA [9] is a hybrid algorithmic-crowdsourcing platform for academic knowledge discovery.
Mixed-initiative systems have been introduced as interactive, intelligent approaches that combine the strengths of human interaction with the algorithmic power of AI to solve problems that could not be solved by either computers or humans alone [11]. In this kind of systems, human work is both used directly to complete HITs and as feedback for machine learning classifiers training to complete that work in the future. Such hybrid approach can be particularly fruitful in crowd science projects to refine machine-extracted metadata while providing evidence on demand by using automatic classification techniques enabled by human crowd workers who can process, filter, classify, and verify the information [10].
3 On the Deployment of SciCrowd: Targeting Crowd-Enabled Scientific Discovery
A paradigmatic shift in information systems design that can be taken into account in the design and development of a crowd-computing system for scientific discovery involves the consideration of IT artifacts and social practices as central units of research [15]. Theoretically grounded from a design science perspective, such approach includes goal-directed collective activities, routines, and resources to create explicit knowledge by means of reflection and concept formation (e.g., linguistic signs, organizational schemes, technical artifacts) as socially situated practices.
Research tasks such as performing a systematic literature review involve a set of laborious, time-consuming, and sometimes wasteful efforts. This includes the identification of relevant sources, data extraction, and quality assessment [5]. In this scenario, relevant sources and data dimensions can be ignored during the process. Thus, suitable crowd-powered systems are required to produce, share, filter, and combine scientific discoveries. In the literature, crowdsourcing has been closely related to processes, task characteristics, and participants [10]. Tinati and colleagues [7] identified numerous factors when designing citizen science platforms, including task specificity, community development, task design, and engagement. The authors pointed out to the likelihood of using periodic feedback, roles and privileges, and task context as motivational factors. In addition, the environment must be easy to use, fast, intuitive, and reliable. Crowd aspects (including size, type, social structure, behavior), quality control, and temporal issues (e.g., daily devoted time) assume particular importance on the design of crowd-powered systems [10]. Hybrid classifiers must be trained according to the crowd behavior and the system must enable the crowd to selectively supplement weak regions of the classifier [16].
3.1 Baseline System Design and Prototyping
SciCrowd’s system prototype.
Currently, the prototype comprises a limited set of features that range from edit a publication record automatically extracted from DBLP3 or added manually by users to the settings panel (e.g., group roles) and main page. Publication details can be visualized pressing the “show” button, which opens an internal page that allows to insert data about a paper. For instance, consider the following scenario. A paper on “medical informatics” could be classified by subarea (e.g., cognitive aging), method, setting, key concepts, findings, and social-technical aspects concerning a certain technology (e.g., a Wiki intended to support knowledge exchange in public health). All these data could be filtered considering specific research purposes. For example, a user could be interested on understanding relations among concepts and related areas in machine learning for medical purposes and/or identifying features in health care systems implemented by Portuguese researchers in a certain period.
System architecture.
3.2 Preliminary Evaluation
The initial results of a survey (N = 90 researchers) corroborate the need for developing a crowd-computing system for research data analysis. The sample consisted mostly of researchers from USA (33.3%) with more than 8 years of experience (60%) and a research focus on crowdsourcing (58.9%). All participants (75.5% male, 24.5% female) were asked to participate via e-mail in a total of 3625 e-mail addresses. Google Forms4 was chosen for collecting responses between September 20, 2017 and April 3, 2018. Researchers pointed out resources (63.3%) and access to information (30%) as the main problems when working on their research tasks. There was some disagreement with regard to the use of crowdsourcing in research activities. Nonetheless, the participants identified classification and categorization (45.6%), verification and validation (44.4%), surveys (40%), content creation (36.7%), and information finding (32.2%) as their main types of tasks assigned to crowd workers. When asked about the incentives for participation in crowd-driven research, researchers pointed out the interest in the topic (31.1%), money and extrinsic rewards (16.7%), social incentives (16.7%), and altruism and meaningful contribution (15.6%). Regarding the best practices and the main drawbacks of using crowdsourcing for scientific purposes, participants identified approaches like gamification, need for supervision, real-time collaboration, proper diffusion, and data collection. Participants also indicated legal and ethical limitations that will make it difficult for people to help collect or analyze research data. Examples include privacy (57.8%), accuracy of information (53.3%), protection of intellectual property (40%), and abuse of personal information (37.8%).
The implementation of SciCrowd as a hybrid approach for integrating human inputs into AI systems was also addressed. Almost all respondents stated that it is “very promising” in the scientific context. A participant called our attention to value human input, while other responses emphasized the importance of interdisciplinary identity, transparency, and provenance “to keep track of all versions of the information and knowledge capture, so that the state of knowledge can be rewound and replayed from any point in time to any other point in time”. Another participant explicitly demonstrated the potential of crowd-computing hybrids by assuming that “machine learning algorithms could learn from user/volunteer responses in applications where the computer classification skill is below a desirable threshold.” A few of the researchers mentioned the risk of enabling machine imposing (wrong) interpretations.
An important implication for design relies on considering diversity (in terms of education and income levels, race, gender) when designing crowd-based human computation systems. Some participants mentioned some problems associated to error tolerance, ethical dilemmas of introducing AI, and crowd bias. An additional set of questionnaire responses showed us some features that can be implemented in SciCrowd. Examples include user-friendly AI algorithms, learn from examples, statistical analysis, a semantic search engine exploring scientific literature by specific questions (instead of keywords), visualization of work progress, a graph of papers based on their impact, and ability to identify the “right crowd” (general crowd might not be suitable for specialized tasks like scientific analysis). In addition, participants also indicated a need for more connections (links) between the data, the published papers, the search (query) terms that led a user to the data, and how the data were used and/or analyzed.
3.3 Limitations and Future Work
Some limitations remain unfilled concerning the lack of case studies to validate the system (scarcity of tests and experimental research) by analyzing interaction logs and questionnaires. An approach for the lightweight development of this system relies on involving the crowd in the underlying engineering and design processes. Design and evaluation principles will be further explored, while usability tests are also required to gain insight on work practices that can better inform the development of a robust, hybrid machine-crowd intelligent system. Based on the requirements identified previously, future deployments include hybrid classifiers and learners trained according to the crowd behavior, a public dashboard, a fine-grained classification mechanism that would store and retrieve metadata that could be extracted in the form of graphics and statistics, and a context-aware reputation system. One important problem in this study was that the survey was conducted with researchers who had not used the system, so results are very limited. A new survey and a controlled experiment engaging users on the evaluation of publication records using a crowdsourcing platform such as MTurk constitute important lines of further research.
4 Concluding Remarks
This paper emphasizes the implications for design concerning a crowd-enabled, self-organizing system under development towards achieving a higher level of engagement by researchers and citizen scientists. SciCrowd system’s prototype aims to improve the way as publication records are evaluated while harnessing the wisdom of a crowd. The main target relies on improving the effective use of data to solve complex problems at large-scale, and the possibility to adapt this tool to multivariate scientific fields. The crowd can be particularly helpful in this process by indexing metadata and fixing errors while providing new interpretations on the published data through highly detailed and filtered data granules. The outputs can be established on understanding how research communities form and evolve over time. Further steps will be focused on providing features for supporting hybrid reasoning using AI inferring on contributions from crowd members performing HITs such as systematic literature reviews and scientometric studies. Despite the prospects for research in crowd science, we must be able to consider incentives, ethical constraints, and social-technical aspects for “closing the loop” in developing crowd-powered scientific computing systems.
Footnotes
Notes
Acknowledgements
This work is financed by the ERDF – European Regional Development Fund through the Operational Programme for Competitiveness and Internationalisation - COMPETE 2020 Programme within project «POCI-01-0145-FEDER-006961», and by National Funds through the Portuguese funding agency, FCT - Fundação para a Ciência e a Tecnologia as part of project «UID/EEA/50014/2013».
References
- 1.Gil, Y., Greaves, M., Hendler, J., Hirsh, H.: Amplify scientific discovery with artificial intelligence. Science 346(6206), 171–172 (2014)CrossRefGoogle Scholar
- 2.Farooq, U., Ganoe, C.H., Carroll, J.M., Giles, C.L.: Designing for e-science: requirements gathering for collaboration in CiteSeer. Int. J. Hum. Comput. Stud. 67, 297–312 (2009)CrossRefGoogle Scholar
- 3.Thelwall, M.: Using altmetrics to support research evaluation. In: Proceedings of the International Workshop on Altmetrics for Research Outputs Measurements and Scholarly Information Management, pp. 11–28 (2018)Google Scholar
- 4.Correia, A., Fonseca, B., Paredes, H.: Exploiting classical bibliometrics of CSCW: classification, evaluation, limitations, and the odds of semantic analytics. In: Proceedings of the International Conference on Human Factors in Computing and Informatics, pp. 137–156 (2013)Google Scholar
- 5.Kitchenham, B., Brereton, O.P., Budgen, D., Turner, M., Bailey, J., Linkman, S.: Systematic literature reviews in software engineering – a systematic literature review. Inf. Softw. Technol. 51(1), 7–15 (2009)CrossRefGoogle Scholar
- 6.Cooper, S., et al.: The challenge of designing scientific discovery games. In: Proceedings of the ACM International Conference on the Foundations of Digital Games, pp. 40–47 (2010)Google Scholar
- 7.Tinati, R., Van Kleek, M., Simperl, E., Luczak-Rösch, M., Simpson, R., Shadbolt, N.: Designing for citizen data analysis: a cross-sectional case study of a multi-domain citizen science platform. In: Proceedings of the ACM CHI Conference on Human Factors in Computing Systems, pp. 4069–4078 (2015)Google Scholar
- 8.Chau, D.H., Kittur, A., Hong, J.I., Faloutsos, C.: Apolo: making sense of large network data by combining rich user interaction and machine learning. In: Proceedings of the ACM CHI Conference on Human Factors in Computing Systems, pp. 167–176 (2011)Google Scholar
- 9.Dong, Z., Lu, J., Ling, T.W., Fan, J., Chen, Y.: Using hybrid algorithmic-crowdsourcing methods for academic knowledge acquisition. Clust. Comput. 20(4), 3629–3641 (2017)CrossRefGoogle Scholar
- 10.Correia, A., Schneider, D., Fonseca, B., Paredes, H.: Crowdsourcing and massively collaborative science: a systematic literature review and mapping study. In: Rodrigues, A., et al. (eds.) CRIWG 2018. LNCS, vol. 11001, pp. 133–154 (2018)Google Scholar
- 11.Horvitz, E.: Principles of mixed-initiative user interfaces. In: Proceedings of the ACM CHI Conference on Human Factors in Computing Systems, pp. 159–166 (1999)Google Scholar
- 12.Bontcheva, K., Roberts, I., Derczynski, L., Rout, D.: The GATE crowdsourcing plugin: crowdsourcing annotated corpora made easy. In: Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pp. 97–100 (2014)Google Scholar
- 13.Borne, K.: Collaborative annotation for scientific data discovery and reuse. Bull. Am. Soc. Inf. Sci. Technol. 39(4), 44–45 (2013)CrossRefGoogle Scholar
- 14.Dunne, C., Shneiderman, B., Gove, R., Klavans, J., Dorr, B.: Rapid understanding of scientific paper collections: integrating statistics, text analytics, and visualization. J. Assoc. Inf. Sci. Technol. 63(12), 2351–2369 (2012)CrossRefGoogle Scholar
- 15.Rohde, M., Stevens, G., Brödner, P., Wulf, V.: Towards a paradigmatic shift in IS: designing for social practice. In: Proceedings of the DESRIST (2009)Google Scholar
- 16.Cheng, J., Bernstein, M.S.: Flock: hybrid crowd-machine learning classifiers. In: Proceedings of the ACM CSCW, pp. 600–611 (2015)Google Scholar
- 17.Zhao, Y., Zhu, Q.: Evaluation on crowdsourcing research: current status and future direction. Inf. Syst. Front. 16, 1–18 (2012)Google Scholar
- 18.Correia, A., Santos, J., Azevedo, D., Paredes, H., Fonseca, B.: Putting “human crowds” in the loop of bibliography evaluation: a collaborative working environment for CSCW publications. Procedia Technol. 9, 573–583 (2013)CrossRefGoogle Scholar
- 19.Börner, K., Scharnhorst, A.: Visual conceptualizations and models of science. J. Informetr. 3(3), 161–172 (2009)CrossRefGoogle Scholar
- 20.Burrows, S., Potthast, M., Stein, B.: Paraphrase acquisition via crowdsourcing and machine learning. ACM Trans. Intell. Syst. Technol. 4(3), 43 (2013)CrossRefGoogle Scholar
- 21.Rzeszotarski, J., Kittur, A.: CrowdScape: interactively visualizing user behavior and output. In: Proceedings of the Annual ACM Symposium on UIST, pp. 55–62 (2012)Google Scholar
- 22.Chilton, L.B., Little, G., Edge, D., Weld, D.S., Landay, J.A.: Cascade: crowdsourcing taxonomy creation. In: Proceedings of the ACM CHI Conference on Human Factors in Computing Systems, pp. 1999–2008 (2013)Google Scholar
- 23.Tchoua, R.B., et al.: Towards a hybrid human-computer scientific information extraction pipeline. In: Proceedings of the IEEE 13th International Conference on e-Science, pp. 109–118 (2017)Google Scholar

