Conference PaperPDF Available

Understanding Science 2.0: Crowdsourcing and Open Innovation in the Scientific Method

Authors:

Abstract and Figures

The innovation process is currently undergoing significant change in many industries. The World Wide Web has created a virtual world of collective intelligence and helped large groups of people connect and collaborate in the innovation process [1]. Von Hippel [2], for instance, states that a large number of users of a given technology will come up with innovative ideas. This process, originating in business, is now also being observed in science. Discussions around "Citizen Science" [3] and "Science 2.0" [4] suggest the same effects are relevant for fundamental research practices. "Crowdsourcing" [5] and "Open Innovation" [6] as well as other names for those paradigms, like Peer Production, Wikinomics, Swarm Intelligence etc., have become buzzwords in recent years. However, serious academic research efforts have also been started in many disciplines. In essence, these buzzwords all describe a form of collective intelligence that is enabled by new technologies, particularly internet connectivity. The focus of most current research on this topic is in the for-profit domain, i.e. organizations willing (and able) to pay large sums to source innovation externally, for instance through innovation contests. Our research is testing the applicability of Crowdsourcing and some techniques from Open Innovation to the scientific method and basic science in a non-profit environment (e.g., a traditional research university). If the tools are found to be useful, this may significantly change how some research tasks are conducted: While large, apriori unknown crowds of "irrational agents" (i.e. humans) are used to support scientists (and teams thereof) in several research tasks through the internet, the usefulness and robustness of these interactions as well as scientifically important factors like quality and validity of research results are tested in a systematic manner. The research is highly interdisciplinary and is done in collaboration with scientists from sociology, psychology, management science, economics, computer science, and artificial intelligence. After a pre-study, extensive data collection has been conducted and the data is currently being analyzed. The paper presents ideas and hypotheses and opens the discussion for further input. © Selection and peer-review under responsibility of FET11 conference organizers and published by Elsevier B.V.
Content may be subject to copyright.
Available online at www.sciencedirect.com
Procedia Computer Science 7 (2011) 327–329
The European Future Technologies Conference and Exhibition 2011
Understanding Science 2.0: Crowdsourcing and Open Innovation in
the Scientific Method
Thierry Bücheler a,, Jan Henrik Sieg b
aArtificial Intelligence Lab, University of Zurich, Andreasstrasse 15, 8050 Zurich, Switzerland
bChair of Strategic Mgmt and Innovation, ETH Zurich, Kreuzplatz 5, 8032 Zurich, Switzerland
Abstract
The innovation process is currently undergoing significant change in many industries. The World Wide Web has created a virtual
world of collective intelligence and helped large groups of people connect and collaborate in the innovation process [1]. Von Hippel
[2], for instance, states that a large number of users of a given technology will come up with innovative ideas. This process, originating
in business, is now also being observed in science. Discussions around “Citizen Science” [3] and “Science 2.0” [4] suggest the same
effects are relevant for fundamental research practices. “Crowdsourcing” [5] and “Open Innovation” [6] as well as other names for
those paradigms, like Peer Production, Wikinomics, Swarm Intelligence etc., have become buzzwords in recent years. However,
serious academic research efforts have also been started in many disciplines. In essence, these buzzwords all describe a form of
collective intelligence that is enabled by new technologies, particularly internet connectivity. The focus of most current research
on this topic is in the for-profit domain, i.e. organizations willing (and able) to pay large sums to source innovation externally, for
instance through innovation contests. Our research is testing the applicability of Crowdsourcing and some techniques from Open
Innovation to the scientific method and basic science in a non-profit environment (e.g., a traditional research university). If the tools
are found to be useful, this may significantly change how some research tasks are conducted: While large, apriori unknown crowds
of “irrational agents” (i.e. humans) are used to support scientists (and teams thereof) in several research tasks through the internet,
the usefulness and robustness of these interactions as well as scientifically important factors like quality and validity of research
results are tested in a systematic manner. The research is highly interdisciplinary and is done in collaboration with scientists from
sociology, psychology, management science, economics, computer science, and artificial intelligence. After a pre-study, extensive
data collection has been conducted and the data is currently being analyzed. The paper presents ideas and hypotheses and opens the
discussion for further input.
© Selection and peer-review under responsibility of FET11 conference organizers and published by Elsevier B.V.
Keywords: Crowdsourcing; Open Innovation; Simulation; Agent-Based Modeling; Science 2.0; Citizen Science
1. Introduction and open questions
Fundamental research is still driven by many thinkers and doers cracking a problem alone, based on their own
knowledge and skills. However, vast exchange, often with participants from different backgrounds in different settings,
takes place in modern research and contemporary research projects are characterized by intense interactions between
groups and individuals, e.g., during idea generation, formulation of hypotheses, evaluation, and data analysis, among
many other research tasks. Large project conglomerates (e.g., EU-funded research or projects funded through the
Advanced Technology Program in the U.S.) actively foster such interactions. In many cases, the scientist groups self-
Corresponding author.
1877-0509/$ – see front matter © Selection and peer-review under responsibility of FET11 conference organizers and published by Elsevier B.V.
doi:10.1016/j.procs.2011.09.014
328 T. Bücheler, J.H. Sieg / Procedia Computer Science 7 (2011) 327–329
Fig. 1. Simplified Research Process, containing “tasks”: the Research Value Chain – from [9].
organize according to their individual strengths and skills (and other reasons) to reach a common goal, without a
strong centralized body of control (see e.g., [7],[8]). Interactions between these individuals and groups can be seen as
instances of collective intelligence, including consensus decision making, mass communication, and other phenomena
(see [9] for further details).
If basic science has become a collective intelligence effort, can it use the ideas and technologies from Crowdsourcing
and Open Innovation to become more efficient and effective relative to the money spent while maintaining the necessary
quality and validity levels? Which are the right incentives to include large groups in basic science and to foster sharing
of ideas and data? Based on empirical data, our research seeks to provide answers to these research questions.
2. Crowdsourcing and Open Innovation
Crowdsourcing and Open Innovation are two terms coined in the last seven years, influencing several research fields.
We use the following two working definitions:
“Crowdsourcing is the act of taking a job traditionally performed by a designated agent (usually an employee) and
outsourcing it to an undefined, generally large group of people in the form of an open call.” [5]
“’Open Innovation’ is a paradigm that assumes that firms can and should use external ideas as well as internal ideas,
and internal and external paths to market, as the firms look to advance their technology. Open Innovation combines
internal and external ideas into architectures and systems whose requirements are defined by a business model.” [6]
3. How to analyze the Scientific Method?
In order to investigate “basic science” in a structured manner, we have simplified the tasks that are conducted in
most scientific inquiries (see Fig. 1) and used MIT’s “Collective Intelligence Gene” framework to analyze the tasks in
combination with the “Three Constituents Principle” from AI [10]. See [9] for details regarding the Simplified Research
Process and the framework used. Based on this categorization and taxonomy, we hypothesize that the following scientific
tasks are especially suited for Crowdsourcing: develop and choose methodology, Identify team of co-workers, gather
information and resources (prior work and implications), analyze data, retest.
4. Where we currently stand
During the ShanghAI Lectures 2009, a global lecture on Artificial Intelligence involving 48 universities from 5
continents, we have collected data to test these hypotheses and to test the other tasks for “crowdsourceability” (as
described in [9]): “The participants supported one of four current scientific projects by contributing a paper stating
their ideas on open questions. Some of the solutions were rated “excellent“, “well-elaborated” and “useful for the
advancement of the project” by the scientists that headed the projects. We sent questionnaires to 372 participating
students after the lectures and received 84 valid replies (23%). Although only 16.0% thereof stated that they had prior
theoretical or technical knowledge regarding the chosen subject, 22.6% of all participants perceived a significant impact
on current research if they participated in the contest. However, initial data collection during this pre-test was insufficient
to analyze all variables in our framework.“In the fall semester 2010, we expanded data collection and rigorously applied
the framework. We also used existing scales from other contexts (e.g., Crowdsourcing in economic environments) to
compare basic science with the corporate R&D domain. This time, we got a reply from 195 participants representing
51 teams (response rate of >70%). 57% of participants had at least a Bachelor’s degree. 44.9% consider themselves
rather specialists in a field (by education) than generalists with broad scientific knowledge. The data is currently being
T. Bücheler, J.H. Sieg / Procedia Computer Science 7 (2011) 327–329 329
analyzed, but a preliminary analysis of the data shows that from the 12 science projects available, representing an almost
complete research process from “define the question” to “interpret data” and “draw conclusions”, all have been chosen
by at least one team (10 teams chose the most popular project, “develop proposal”). 41.5% of the participants were at
least “satisfied” (score 5 or better in a 7-points Likert scale) by this experience and 57% are positive about working on
such a Crowdsourcing project in science again. This last dimension (not very surprisingly) correlates strongly with the
fun level that each participant perceived. The participants had zero financial incentives. The only thing they could get
by delivering a solution to the projects were 3 credit points (out of 61 necessary) to pass a lecture. Only 22% indicated
that they would have put more effort into their project if money had been awarded to the best solution (score 5 or higher
of 7). This group is almost disjoint from the group that was satisfied with the project at a score of 5 or more (from 7).
Again, the researchers supervising the projects were positively surprised by the quality, accuracy, and usefulness of
the results.
5. What we want to do next
The research team will now thoroughly analyze the data gathered in this second round of data collection. In
parallel, the team has started to implement a simulator for testing the identified local rules of interaction in such a
Crowdsourcing/Open Innovation context and other findings, comparing them with empirical data from other disciplines
(e.g., management science). In addition, this simulator allows us to better understand sensitivities of parameters that
researchers can set/influence and therefore might have some predictive power.
6. How you can be part of this
The team is reaching out to partners from other scientific domains (e.g., psychology for teamwork and brainstorming,
biology for consensus decision making, swarm behavior, etc.). If you believe you have a connection or interesting idea
that fits this topic or if you would like to challenge the ideas presented here, please get in contact with the authors.
References
[1] T.W. Malone, R. Laubacher, C. Dellarocas, Harnessing Crowds: Mapping the Genome of Collective Intelligence, in: MIT Sloan Research
Paper, 4732-09, 2009.
[2] E. von Hippel, Democratizing innovation, MIT Press, Cambridge, Mass, 2005.
[3] A. Irwin, Citizen science. A study of people, expertise and sustainable development Environment and society, Routledge, London, 1995.
[4] B. Shneiderman, Science 2.0 Copernican challenges face those who suggest that collaboration, not computation are the driving energy for
socio-technical systems that characterize Web 2.0, Science 319 (2008) 1349–1350.
[5] Howe, J. 2010. Crowdsourcing. Why the Power of the Crowd is Driving the Future of Business. http://www.crowdsourcing.com/. Accessed 20
Feb. 2011.
[6] H.W. Chesbrough, Open innovation. The new imperative for creating and profiting from technology, Harvard Business School Press, Boston,
Mass, 2003.
[7] G. Melin, Pragmatism and self-organization Research collaboration on the individual level, Research Policy 29 (1) (2000) 31–40.
[8] K. Stoehr, WHO, A multicentre col8laboration to investigate the cause of severe acute respiratory syndrome, The Lancet 361 (9370) (2003)
1730–1733.
[9] T. Buecheler, J.H. Sieg, R.M. Füchslin, R. Pfeifer, Crowdsourcing Open Innovation and Collective Intelligence in the Scientific Method: A
Research Agenda and Operational Framework, in: H. Fellermann, et al. (Eds.), Artificial Life XII. Proceedings of the Twelfth International
Conference on the Synthesis and Simulation of Living Systems, MIT Press, Cambridge, Mass, 2010, pp. 679–686.
[10] R. Pfeifer, J. Bongard, How the body shapes the way we think A new view of intelligence. A Bradford book, MIT Press, Cambridge, Mass,
2007.
... In addition, a group of publications with an "Education" orientation focuses on libraries' support for "science 2.0" (see, e.g., [97]) and its sustainability dimension, libraries' involvement in the development of sustainability-related curricula, as well as ways to raise awareness and "information literacy" levels of sustainability knowledge among students. ...
Article
Full-text available
Sustainability issues constitute a distinct subdiscipline of librarianship and information science, with its own areas of study, methods, and areas of application. Despite being nearly 30 years old, there are still divergent opinions on its current phase of development and its links to other scientific disciplines. The authors aim to clarify and summarize the ongoing discussion through citation analysis, shedding light on the lifecycle of research papers in sustainability-oriented library and information science, the current research subjects of focus, the influence of subdomains within the discipline on other scientific areas, and the overall quantitative and qualitative landscape of the discipline. A detailed elucidation of the inquiry’s results is intended to outline the discipline’s cognitive structure and its impact on sustainability science. The lifecycle of disciplinary papers indicates the dynamic development of the field. Sustainability-oriented library and information science is well-established, and its research focus has already been consolidated. The optimal citation window for measuring the impact strength in this discipline is a period of 3 to 4 years. “Culture” and “Education” have been identified as the most forward-looking subdisciplines, whereas “Buildings” and “Collections” exhibit less dynamic growth. The social sustainability pillar is the dominant one, while the environmental pillar is slightly less prominent. The economic pillar is the least represented. Although the majority of information exchange occurs within the discipline, it maintains strong and numerous links with several other fields, including both technical and social sciences, as well as the humanities.
... Although previous publications contribute to the recognition and understanding of crowdsourcing in science (Beck et al. 2022;Bücheler and Sieg 2011), 'the use of crowdsourcing in science is still being tested' (Bassi et al. 2020, 302). It is increasingly being noticed that many factors may limit the use of crowdsourcing in science (Law et al. 2017;Schlagwein and Daneshgar 2014). ...
... However, recently, we have witnessed the emergence of other models, starting from Science 2.0 to Science 6.0, which put the emphasis on larger interdisciplinarity, participation, and collaboration, paying more attention to individual and social factors and indicating the directions of scientific development in the digital age. Thus, for example, Science 2.0 denotes the use of digital technologies and the Internet in scientific research, such as data sharing and scientific collaboration, thus promoting the reproducibility of research results and the speed of the research process [133]. Science 3.0 reflects greater public participation and collaboration within scientific research to ensure the conformity of research with the needs of society [134], while Science 4.0. ...
Article
Full-text available
In a contemporary world facing countless multifaceted crises and challenges, science can still serve as one of the most powerful tools to deal with the ordeals of our time. However, the scientific community needs to provide space for reflection on novel ways of developing its centuries-old heritage and unlocking its potential for the benefit of the world and humanity. The purpose of this article was to deliberate on the image of contemporary science within the framework of the new philosophical paradigm of metamodernism. Following historical strands related to metamodernism and science, the authors encircled the general features and elaborated the main philosophical principles of metamodernism. The main task was to identify elements of contemporary science that conform to the philosophical principles of metamodernism. Thus, several features of science and research, such as the structure of science, scientific truth, metanarratives of science, scientific thinking, system of science, interaction of scientific disciplines, dialogue of science with society and politics, open science, digitalisation of science, etc., were interpreted through the perspective of the ontological, epistemological, axiological, and methodological principles of metamodernism. This article ends with a summary of the main points of the discussion and practical implications of the presented ideas.
... La ciencia 2.0 [47,48] es la aplicación de las tecnologías de la web social al proceso científico. La web social, web 2.0 o web participativa se caracteriza por el empleo de tecnologías abiertas, tanto desde el punto de vista de la arquitectura de la información, como de la interconexión de servicios y, sobre todo, del trabajo colectivo que se realiza de forma telemática, colaborativa y desinteresada. ...
Article
Full-text available
En este manuscrito presentamos una aplicación web con soporte en lenguaje de programación Python ReactJS y JavaScript, libre y abierta, para el desarrollo de una actividad de enseñanza-aprendizaje de la astronomía, específicamente para el cálculo de la rotación diferencial del Sol para estudiantes y público en general en edad escolar entre 10 y 18 años. El propósito fundamental es la de difundir el conocimiento del Sol y algunas de sus propiedades. La aplicación web es autocontenida y con suficiente guía y ayuda para que cualquiera pueda usarla, además de su dinamismo y diseño innovador, pretende presentar estrategias agradables para la enseñanza y aprendizaje de la ciencia en torno al Sol.
Article
Full-text available
Operacjonalizacja problemu badawczego stanowi konstytutywny etap procesu realizacji każdego badania naukowego. Jednak jej przeprowadzenie jest dla badacza wyzwaniem, wymaga bowiem dobrej, pełnej i aktualnej znajomości analizowanej problematyki. Dlatego też coraz częściej zachęca się badaczy do poszukiwania sposobów czy postępowań, dzięki którym możliwe jest doprecyzowanie i zrozumienie w sposób wieloaspektowy zjawisk, które mogą stać się przedmiotem badań empirycznych. W szczególności nabiera to znaczenia w kontekście badań nad szkolnictwem wyższym, które wymagają podejścia jakościowego. Crowdsourcing naukowy wszedł w fazę popularyzacji i wydaje się obiecujący w kontekście operacjonalizacji problemu badawczego. Celem artykułu jest przedstawienie propozycji wykorzystania crowdsourcingu naukowego jako pomocniczego (uzupełniającego) postępowania na etapie operacjonalizacji problemu badawczego oraz zastanowienie się nad jego skutecznością w kontekście badań nad szkolnictwem wyższym.
Article
Full-text available
Introduction Digital health and evolutionary medicine create new insights of mediation and health treatment plan support, introducing crowdsourcing and patients’ real-world data records, so as to promote the development of high-quality healthcare accessible to everyone. Within the scope of its activities Metabio’s team has developed an interoperable unified method and technology for crowd-generated databases, creating a user-friendly platform for data collection, processing, and distribution among stakeholders within the global healthcare system in real time. Methods In this paper we describe standard methodologies, requirements, issues, and challenges for the design and deployment of an advanced IT infrastructure for longitudinal structured patient-related data records, based on a patient-centric model of operation, as well as the difficulties for the development of disease-specific user-prefixed interface for real-world data collection. Results Through a dynamic real-time (DRT) e-consent module and digital rights management protocols, the overall platform enables patients to monitor and manage their disease-related conditions, as well as for healthcare providers and/or research entities to have access to valuable biomedical patient data, not recorded so far. Conclusion The project introduces novel perspectives for future evidence-based practices, promoting research and development and improving current healthcare systems, by using crowd-generated data sources that bring a much higher degree of accuracy and value for the entire healthcare system.
Article
Full-text available
The lonely researcher trying to crack a problem in her office still plays an important role in fundamental research. However, a vast exchange, often with participants from different fields is taking place in modern research activities and projects. In the "Research Value Chain" (a simplified depiction of the Scientific Method as a process used for the analyses in this paper), interactions between researchers and other individuals (intentional or not) within or outside their respective institutions can be regarded as occurrences of Collective Intelligence. "Crowdsourcing" (Howe 2006) is a special case of such Collective Intelligence. It leverages the wisdom of crowds (Surowiecki 2004) and is already changing the way groups of people produce knowledge, generate ideas and make them actionable. A very famous example of a Crowdsourcing outcome is the distributed encyclopedia "Wikipedia". Published research agendas are asking how techniques addressing "the crowd" can be applied to non-profit environments, namely universities, and fundamental research in general. This paper discusses how the non-profit "Research Value Chain" can potentially benefit from Crowdsourcing. Further, a research agenda is proposed that investigates a) the applicability of Crowdsourcing to fundamental science and b) the impact of distributed agent principles from Artificial Intelligence research on the robustness of Crowdsourcing. Insights and methods from different research fields will be combined, such as complex networks, spatially embedded interacting agents or swarms and dynamic networks. Although the ideas in this paper essentially outline a research agenda, preliminary data from two pilot studies show that non-scientists can support scientific projects with high quality contributions. Intrinsic motivators (such as "fun") are present, which suggests individuals are not (only) contributing to such projects with a view to large monetary rewards.
Article
What goes on in the scientific networks and the research teams? What is the collaborative situation like? Why do scientists collaborate? This study focuses on the micro level of research collaboration and investigates the reasons for and effects of collaboration for the individual scientist through a survey and a number of interviews. The interaction within the research team is highlighted, showing the feelings and conditions which encompass the teamwork. The empirical findings are conceptualized in a model where research collaboration is suggested to be understood as dependent on how the reasons, forms and effects respectively vary. The collaborations are characterized by strong pragmatism and a high degree of self-organization. Finally, the science policy implications of this study are discussed. It is suggested that research policy should provide financial and organizational possibilities for the researchers to establish joint ventures and also fund projects on a team or network basis.
Article
Severe Acute respiratory syndrome is a new disease in human beings, first recognised in late February, 2003, in Hanoi, Vietnam. The severity of the disease, combined with its rapid spread along international air-travel routes, prompted WHO to set up a network of scientists from 11 laboratories around the world to try to identify the causal agent and develop a diagnostic test. The network unites laboratories with different methods and capacities to rapidly fulfil all postulates for establishing a virus as the cause of a disease. Results are shared in real time via a secure website, on which microscopy pictures, protocols for testing, and PCR primer sequences are also posted. Findings are discussed in daily teleconferences. Progress is further facilitated through sharing between laboratories of samples and test materials. The network has identified a new coronavirus, consistently detected in samples of SARS patients from several countries, and conclusively named it as the causative agent of SARS; the strain is unlike any other known member of the genus Coronavirus. Three diagnostic tests are now available, but all have limitations.
Book
Annotationnewline newline Annotation.newline newline Annotation
Article
Over the past decade, the rise of the Internet has enabled the emergence of surprising new forms of collective intelligence. Examples include Google, Wikipedia, Threadless, and many others. To take advantage of the possibilities these new systems represent, it is necessary to go beyond just seeing them as a fuzzy collection of “cool” ideas. What is needed is a deeper understanding of how these systems work. This article offers a new framework to help provide that understanding. It identifies the underlying building blocks—to use a biological metaphor, the “genes”—at the heart of collective intelligence systems. These genes are defined by the answers to two pairs of key questions: – Who is performing the task? Why are they doing it? – What is being accomplished? How is it being done? The paper goes on to list the genes of collective intelligence—the possible answers to these key questions—and shows how combinations of genes comprise a “genome” that characterizes each collective intelligence system. In addition, the paper describes the conditions under which each gene is useful and the possibilities for combining and re-combining these genes to harness crowds effectively. Using this framework, managers can systematically consider many possible combinations of genes as they seek to develop new collective intelligence systems. ∗ University of Maryland 1
Article
What goes on in the scientific networks and the research teams? What is the collaborative situation like? Why do scientists collaborate? This study focuses on the micro level of research collaboration and investigates the reasons for and effects of collaboration for the individual scientist through a survey and a number of interviews. The interaction within the research team is highlighted, showing the feelings and conditions which encompass the teamwork. The empirical findings are conceptualized in a model where research collaboration is suggested to be understood as dependent on how the reasons, forms and effects respectively vary. The collaborations are characterized by strong pragmatism and a high degree of self-organization. Finally, the science policy implications of this study are discussed. It is suggested that research policy should provide financial and organizational possibilities for the researchers to establish joint ventures and also fund projects on a team or network basis.