Introduction

Confronted with increasingly complex challenges in a fast-changing environment, organizations have to continuously innovate to remain competitive. During the innovation processes, confidence is built through the use of Demonstration Environments (DEs) that offer space for prototyping, testing, and verification of new products, services and systems. DEs, including testbeds, pilot-plants, and living labs, play a crucial role in ensuring that innovative outcomes can effectively fulfill their intended functions (Högman & Johannesson, 2013). These environments are supported and utilized by a diverse set of stakeholders, including but not limited to, universities, government organizations, and private businesses. Accurate differentiation of DEs is essential for selecting and utilizing the most appropriate environment based on their distinct capabilities and characteristics, to achieve specific demonstration outcomes. These outcomes may include cost reduction, improved user experience, and more.

Despite the significance of DEs in practice as well as the growing scholarly literature, there remain notable gaps and challenges in the existing studies pertaining to their differentiation. The need for further research in this area can be attributed to two key factors: (1) the heterogeneity of DE domain knowledge, and (2) the lack of a holistic view of different types of DEs. Consequently, this paper seeks to address the identified gaps by offering a differentiated understanding of DEs, fostering increased knowledge sharing across application fields, and resolving the existing heterogeneity in terminology.

Aligned with the works of Costa et al. (2016), Järvenpää et al. (2019), Montero Jiménez et al. (2021), Chen et al. (2021), the literature gaps can be closed by creating a DE ontology that explicitly structures the entities of a domain and their relationships. This ontology would explicitly structure the domain entities and their relationships, facilitating a comprehensive characterization of DEs that encompasses various types, distinctive features, applications, and involved components, such as stakeholders and facilities (Turchet et al., 2020). Furthermore, the ontology would enable reasoning based on semantic rules, a crucial capability not yet explored in any previous study related to DEs. Not only has no previous study created an ontology for DEs, but this holds for both a range of different DEs and specific DE types, such as testbeds.

This paper aims to address the research question, "How can Demonstration Environments be defined and characterized?" To achieve this, we propose the Demonstration Environments Ontology (DEO) as a representation of domain knowledge related to DEs. The DEO is publicly available at http://purl.archive.org/demonstration-environments-ontology/documentation. Following the METHONTOLOGY methodological framework (Fernández-López et al., 1997), a systematic collection of peer-reviewed English journal articles that specifically focused on Demonstration Environments (DEs) was carried out to develop the ontology. To establish a semantic framework, word embeddings and knowledge extraction through four algorithms to extract domain words, identify concepts and properties, build ontology hierarchy and identify concept-property relations were utilized in the analysis (Mahmoud et al., 2018; Albukhitan & Helmy, 2016). In line with Montero Jiménez et al. (2021) we then manually populated the ontology in Protégé. Indeed, a key contribution of this research is the provision of a systematic framework, realized through the development of the ontology, which fosters a common understanding of DEs and standardizes terminologies used in both literature and practice.

Following a comprehensive review of existing innovation-related ontologies (“Literature review”) and “Methodology” sections of this paper delves into the ontology development methodology and deep-learning steps employed. “Demonstration environments ontology” section presents the main components and structures of the DEO, establishing a robust and standardized framework for understanding DEs. In “Application of DEO” section, the applicability of the DEO is effectively demonstrated through the examination of two use cases. The key findings are discussed in “Discussion” section, affirming the fulfilment of the ontology requirements. The paper concludes with an outline of future research directions in “Conclusion” section.

Literature review

Demonstration environments

DEs can be used in different contexts and for diverse purposes in various stages of the innovation process (Ballon et al., 2005). For example, testbeds are built in an isolated laboratory environment and are commonly seen in the early stage of design and development such as proof-of-concept evaluations (Papadopoulos et al., 2017). Pilot-plants mirror real-life manufacturing plants and they are usually used before commercialization to demonstrate the feasibility of technologies (Hellsmark et al., 2016). Living labs closely resemble the context of the product or service in the real-life environment and involve users as co-creators of innovation outcomes (Almirall et al., 2012). However, current literature lacks a holistic view of different types of DEs. Numerous studies on DEs such as testbeds (Abuarqoub et al., 2012; Kim et al., 2020), pilot-plants (Deiana et al., 2017; Carrera et al., 2022) and living labs (Leminen et al., 2015; Voytenko et al., 2016; De Vita & De Vita, 2021) have shown how these DEs can be used in innovation processes realizing different objectives. While these studies add important insights about a particular DE type, they fail to depict the diverse landscape of DEs and do not offer an integrated perspective on the characteristics and applicability of such environments in different application contexts. Only a few studies have provided holistic insights into the diverse landscape of DEs. The first effort towards the discussion and differentiation of different DEs is reported in Ballon et al. (2005). The authors characterize and benchmark five innovation platforms for broadband innovation, including the prototyping platform, testbed, field trial, living lab, market pilot and societal pilot. However, a limitation of the developed framework is that the authors adopted a general focus on those open and joint platforms only. More importantly, the developed framework is structured as a typology, this prevents the knowledge of DEs from being expanded and joining the ecosystem formed around the innovation. This is important for enriching the domain knowledge of DEs in the future. For example, integrating the domain knowledge of DEs with other innovation-related concepts for better innovation management. Similarly, Schuurman and Tõnurist (2017) carry out a comparison between innovation labs and living labs. They explore similarities and differences between the two DEs and propose a model for further integration of them. The clear limitation is that the study only considers two similar environments. Therefore, current efforts to provide a systematic overview of DEs are limited. This gap leads to the fact that different DEs are not defined in a way such that they can be compared and benchmarked against each other. This made the differentiation between DEs challenging since there is no universal standard for a systematic DE description, annotation and classification. Moreover, the diversity of actors involved in the DEs further adds to the complexity of understanding DEs and how they can be leveraged effectively.

Although recognizing the value of DEs in manufacturing, scanty literature has focused on differentiating DEs for better implementation and application of the DEs in manufacturing. Firstly, there is an inconsistency in the use of terminologies that represent different types of DEs, leading to the heterogeneity of DE domain knowledge. For example, both ‘Living Lab’ and ‘Innovation Lab’ have been used to describe DEs that are based on the concept of open innovation and co-creation involving end-users in the process of innovation (Bergvall-Kareborn & Stahlbrost, 2009; Leminen et al., 2012; Whicher & Crick, 2019; Zurbriggen & Lago, 2019; Bulkeley et al., 2019; Fecher et al., 2020; Greve et al., 2020, 2021). Nevertheless, the two terminologies are not always used consistently in extant literature. Some researchers believe the two terms represent different environments. Schuurman and Tõnurist (2017) argue that living labs and innovation labs are two distinctly different environments. Osorio et al. (2019) also regard the two as different environments. However, other researchers treat innovation labs and living labs as interchangeable terms (Criado et al., 2021). Therefore, current studies on DEs do not share a common understanding of the use of terminologies to represent different DEs. This inconsistency in the use of terminologies indicates an information heterogeneity in defining and characterizing DEs. Such heterogeneity makes it challenging to establish a universal understanding to differentiate DEs in DE-related studies.

Ontology

Ontology is an explicit specification of a conceptualization which refers to an abstract model of some phenomenon that is formed by relevant concepts (Studer et al., 1998). It describes concepts and relationships among them, it can be defined as a 4-tuple: \(\sigma = <C,R,I,A>\),where C is a set of classes (or concepts), R is a set of taxonomic and non-taxonomic relations, I is a set of individuals and A is a set of axioms (Gruber, 1993). Ontology-based semantic technologies are about the activities, theories and principles governing the domain, they can assist in managing, standardizing and integrating heterogeneous information (Fernández-López et al., 1997; El Bassiti & Ajhoun, 2014). In the following parts, we reviewed (1) ontology engineering and its methodologies; (2) ontology learning (OL); and (3) ontology validation approaches.

Ontology engineering methodologies

Ontology engineering investigates a series of principles, methods and tools for activities like knowledge representation, ontology design, standardization, reuse and knowledge sharing (Mizoguchi & Ikeda, 1997). An ontology engineering methodology provides a set of guidelines and activities for ontology development. Several ontology engineering methodologies have been proposed to date. While some methodologies for developing ontologies were proposed at the outset, others emerged as a result of the experience and insights gained during the development of ontologies for various projects.

Based on the experience of developing an ontology in the domain of chemicals, the METHONTOLOGY methodological framework was proposed for developing an ontology for a domain from scratch (Fernández-López et al., 1997). This framework introduces detailed development activities and techniques, including (1) the specification which identifies the audience, scope, use scenarios and requirements; (2) the conceptualization where the domain knowledge will be structured in a conceptual model in terms of the domain vocabulary; (3) the formalization of the ontology from the conceptual model; (4) the integration with existing ontologies; (5) the implementation of ontology with an appropriate environment; and (6) the maintenance of the implemented ontology. Orthogonal to the development activities, there are three support activities to be accomplished during the lifetime of the ontology: knowledge acquisition, documentation and evaluation of the ontology developed (Fernández-López et al., 1997). This methodological framework has been tested through developing ontologies for different domains like music (Turchet et al., 2020), biology (Ayadi et al., 2019b) and manufacturing (Catalano et al., 2009; Talhi et al., 2019).

Ontology Development 101 is another well-known domain ontology development methodology developed by Noy and McGuinness (2001). Also deriving from the experience, the authors provided a detailed example of constructing a wine ontology from scratch. It includes seven stages: (1) Determine the domain and scope of the ontology; (2) Consider reusing existing ontologies; (3) Enumerate important terms in the ontology; (4) Define the classes and class hierarchy; (5) Define the property of classes; (6) Define the facets of the slots; and (7) Create instances for classes. Even though the method includes detailed ontology construction steps, it does not consider the ontology development life cycle.

Besides, editing tools are required for ontology engineering, for example, the Protégé editor (Porzel and Malaka, 2000) which is one of the most prevalent ontology editors, NeOn ToolkitFootnote 1 which is suitable for heavy-weight projects, and SWOOP (Kalyanpur et al., 2006) which is a small and simple ontology editor.

Ontology learning

An ontology can be constructed manually by domain experts and knowledge engineers; or cooperatively where the construction system is supervised by the expert; or (semi-) automatically where there is limited human intervention in the process of construction (Al-Aswadi et al., 2019). Among these methods, a semi-automatic approach is preferred. It not only enables discovering ontological knowledge from large datasets at a faster pace and minimal search efforts but also alleviates the risks of having human-introduced biases and inconsistencies when acquiring the knowledge (Zhou, 2007).

Ontology construction through learning is a semi-automatic approach to creating a new ontology with minimal human effort. Ontology learning is essentially the process of extracting high-level concepts, relations, and sometimes axioms from information in order to construct an ontology. Many established techniques have contributed to the progress in ontology learning, for example, natural language processing, data mining and machine learning (Wong et al., 2012). In recent years, there has been a significant move towards the use of deep learning techniques for ontology construction (Al-Aswadi et al., 2019). Albukhitan and Helmy (2016) introduce a framework that used word embedding models such as Continuous Bag of Words (CBOW) and Skip-gram (SKIP-G) (Mikolov et al., 2013) to extract the taxonomic relation extraction from Arabic text data. Similarly, Mahmoud et al. (2018) proposes an OL methodology that combines word embeddings using Word2Vec (Mikolov et al., 2013) with Balanced Iterative Reducing and Clustering using Hierarchies (BIRCH) (Zhang et al., 1996) algorithm for text big data extraction.

Ontology validation approaches

Ontology validation is important for checking ontology quality. Ayadi et al. (2019b) summarize and evaluate common validation approaches:

  1. 1.

    Task-based approach: This approach applies ontology to certain tasks and evaluates performance (Porzel & Malaka, 2004). However, it does not assess the ontology structure and tends to ignore the shortcomings in ontology conceptualization.

  2. 2.

    Automated consistency checking: This approach uses a Description Logic reasoner like HermiT to check the internal consistency of an ontology, for example, if the ontology contains any contradictory information (Turhan, 2011). However, this approach ignores the background knowledge of ontology.

  3. 3.

    Gold standard checking: This approach compares the constructed ontology with a benchmark ontology by measuring the conceptual and lexical similarities (Maedche & Staab, 2002). One limitation is that there might be a lack of domain ontology for benchmarking.

  4. 4.

    Criteria-based approach: This approach evaluates the ontology with a set of predefined criteria (Table 1) and it combines both quantitative and qualitative methods of validation (Vrandečić 2009). Nevertheless, some of the criteria validation can be highly dependent on subjective judgement.

  5. 5.

    Data-driven validation: This approach compares the ontology with the domain data to be covered by the ontology, but the correctness and clarity cannot be verified with this approach (Brewster et al., 2004).

  6. 6.

    Expert validation: This approach evaluates the ontology against a set of predefined criteria, requirements and standards by expert judgements, but the validation result is subjective and depends on the capability of the experts (Lozano-Tello & Gomez-Perez, 2004). Indeed, the ontology validation approach should be selected based on the validation purposes, ontology application of which an ontology is to be used, and aspects that an ontology is to be validated or evaluated (Walisadeera et al., 2016).

Table 1 Ontology validation criteria

Innovation-related ontologies

Over the last decades, ontologies gained popularity as a supporting tool for innovation management modelling. The scope of these innovation-related ontologies is wide-ranging, varying from a focused area on innovation idea selection (Bullinger, 2008), to the entire innovation management life-cycle in general (Westerski et al., 2010). These ontologies also differ in their intended purposes and level of abstraction. For example, the OntoGate Ontology (Bullinger, 2008) only covers the early stage of innovation, whereas GI2MO (Westerski et al., 2010) includes the entire innovation management life-cycle. They are also designed for different usages: OntoGate Ontology was developed for idea assessment and selection, while Ontology for Innovation (Greenly, 2012) focuses on ensuring the needs of the idea have been met. Nevertheless, to our knowledge, there is no common ontology created explicitly for DEs, neither individual type nor as a whole.

Ning et al. (2006) introduce the Semantic Innovation Management system that includes an Iteams ontology to create a semantic web of innovation knowledge. The Iteams ontology aims to manage innovation information and to operate innovation processes across the extended enterprise seamlessly. It specifies five abstract classes: Goals, Actions, Results, Teams and Community. This ontology can be used to facilitate the distributed collection and development of ideas.

OntoGate ontology (Bullinger, 2008) is a formal domain heavyweight ontology for the early stage of innovation. This ontology was developed based on empirical research to assist idea assessment and selection. It covers three main aspects for the idea to be assessed: market, strategy and technology. Through introducing four key elements: Participant, Input, Gate and Output, the OntoGate ontology structures a company’s understanding of the innovation process.

The Idea ontology (Riedl et al., 2011) is a lightweight application ontology that supports the innovation life-cycle management in open-innovation scenarios. It provides technical means to represent complex idea evaluations along with various concepts. Unlike the Iteams ontology (Ning et al., 2006), the Idea ontology takes a modular approach which allows it to be complemented by additional concepts like customized evaluation methods.

Fig. 1
figure 1

Research overarching approach, process, tool and method

The Generic Idea and Innovation Management Ontology (Gi2MO) ontology (Westerski et al., 2010) is a domain ontology for idea management systems in innovation. It covers the basic idea management processes including idea generation, idea improvement, idea assessment, idea implementation and idea deployment. This ontology is designed to annotate and describe resources gathered inside idea management facilities.

Zhang et al. (2011) design a multi-layered ontology called the Science and Technology Innovation Concept Knowledge-Base (STICK) ontology to capture important properties of the innovation Eco-system over time. It is comprised of three layers: (1) the basic layer includes innovation concepts and products, which distinguishes the concept of innovation from the materialization of innovation; (2) the extended layer covers innovations, key players, and their dynamic interactions; and (3) an evidence layer contains mainly supporting documents. The STICK ontology focuses on innovations and the individual or organizational members of the innovation communities, and their relationships over time.

Unlike the aforementioned ontologies, Greenly (2012) proposes an ontology to facilitate the matching of needs and innovations. The key class Innovation is related to eight classes including the Development stage, Innovator, Embodiment, Improvement, Benefit, Disruption, Problem, and Need.

Zanni-Merk et al. (2013) present a formal ontology for the Theory of Resolution of Inventive Problems (TRIZ) which is a systematic methodology that helps guide the search for inventive solutions to understanding and solving challenging problems. The TRIZ ontology is developed for creativity assistance, it covers three sub-ontologies:(1) the sub-ontology of the TRIZ system model; (2) the substance-field models sub-ontology; and (3) the contradictions and parameters sub-ontology. The TRIZ ontology stresses more on the formulation of problems and guidance to the solution designs, rather than a later stage of the solution validation or demonstration.

Creative Workshop Management Ontology (CWMO) (Gabriel et al., 2019) is created for a different perspective of creativity management. It describes the area of knowledge of creative workshops to assist people involved in the creative workshop in realizing the problem-solving task. CWMO covers the concepts of evaluation strategy (e.g., “quick evaluation”), evaluation techniques (e.g., “multi-criteria evaluation”) and evaluation criteria as part of the creative workshop process. However, the knowledge of evaluation is limited because this ontology is only for application in the context of the creative workshops to solve the problem collectively. Besides, there are also a few creativity management-related ontologies available, for example, the ontology for e-brainstorming (Lorenzo et al., 2021), linked data source for expert finding (Stankovic, 2010), and engineering knowledge management (Brandt et al., 2008). However, none of them has stressed the demonstration, evaluation or validation of the creative subject.

As described, none of the ontology has explicitly mentioned the demonstration of the idea or innovation during the innovation development. In fact, all these ontologies centered on the innovation or idea itself. Therefore, a new ontology entitled “Demonstration Environment Ontology” (DEO) for the representation of this domain is proposed in this study.

Methodology

The main aim of this study is to provide a formal and semantic representation of DEs as a standardized framework. To accomplish this objective, this study aims to create a domain ontology for DEs. The methodology for designing and developing the DEO followed METHONTOLOGY methodological framework (Fernández-López et al., 1997). Figure 1 details the overarching approach following the METHONTOLOGY framework, ontology development processes and corresponding tools and methods for each process. We have documented our research data in the OSF platform which includes detailed steps, relevant codes and sample data, this is publicly available at https://osf.io/y9nem/.

Specification

Goals, target uses and audience

The main goal of this ontology is to provide a formal domain model for the DEs with definitions and characterizations. There are several intended uses of this ontology:

  1. 1.

    A knowledge representation of the DEs

  2. 2.

    A formal reference model that standardizes the terminologies related to DEs

  3. 3.

    A conceptual model capable of guiding target users in the demonstration process

The target audience of the DEO is represented by all actors and stakeholders that are involved in the innovation life-cycle, including research institutions, incubators & accelerators, startups & enterprises, private companies, governments and so on.

Requirements

The ontology should be designed in a way such that it meets the following requirements (Järvenpää et al., 2019):

  1. 1.

    Providing a structure to define and differentiate DEs by determining aspects that characterize different DEs

  2. 2.

    Ensuring the semantics are clearly defined such that there is no ambiguity in terminology and the meanings in the knowledge representation are clear; redundancy in relevant terminologies has to be captured by this ontology

  3. 3.

    Offering means for searching and querying information and supporting activities such as searching for and selecting suitable DEs and proposing feasible DE designs

  4. 4.

    Allowing the semantic model to be understood and interpreted by different users, applications and systems

Ontology development

Following Fig. 1, this section describes the detailed development processes of the DEO, including (1) Data acquisition; (2) Data pre-processing and training of word embeddings; and (3) Ontology construction.

Data acquisition

This study discovered relevant ontological knowledge from text data which are among the commonly used data sources for constructing the ontology (Faria et al., 2014; Arguello Casteleiro et al., 2017; Mahmoud et al., 2018; Ayadi et al., 2019b; Al-Aswadi et al., 2019; Deepa & Vigneshwari, 2022). To be more specific, this study used peer-reviewed journal articles for constructing the ontology where data diversity, quality and statistical significance can be achieved (Mahmoud et al., 2018; Ayadi et al., 2019a; Kumar & Starly, 2021).

Table 2 Search String related to DEs (Adapted from Greve & O’Sullivan (2019))

To prepare the text corpus for ontology construction, a search string that is relevant to DEs (Table 2) was used to locate peer-reviewed English journal articles on the topic of DEs in the Web of Science (WoS) database. Due to its wide coverage and relevance for this study, WoS was used to retrieve articles. As opposed to, for example, Scopus, WoS indexes also the Technology Innovation Management (TIM) Review, which is the journal with the largest number of special issues and articles on living labs to date (Westerlund et al., 2018). Moreover, WoS has already been employed to analyse similar domains, such as open innovation (Dahlander & Gann, 2010), technology business incubation (Mian et al., 2016), living labs (Greve et al., 2021).

The initial search was conducted on 7th June 2020, which produced a total of 4922 unique articles. Obtained articles were screened with a pre-formulated retrieval strategy. The exclusion criteria comprised: (1) search keywords were included as false-positive words. For example, Evans and Karvonen (2014) use ‘living lab’ as a metaphor to describe a lab with living components; and (2) the research does not discuss DEs. For example, Clark et al. (2018) use ‘testbed’ as an official name of a program (Hazardous Weather Testbed Spring Forecasting Experiment) without discussing the testbed itself. Detailed screening steps can be found in Appendix A, we also detailed the list of false positive words in Table 4. Following the exclusion of irrelevant results, a total of 3621 articles were finalized as the text corpus for further extraction.

To build up and formulate the text corpus for performing the word embedding, the full text of these articles was extracted using tools such as Elsevier’s API, PDFMinerFootnote 2 (20191125 version) to obtain the article in plain text format and store them as separate.txt documents for future analysis.

Data pre-processing and training of word embeddings

This study exploited word embeddings to automate the process of extracting ontology components from relevant journal articles.

This process consists of three steps: (1) pre-processing the text acquired; and (2) word embedding training and verification. These steps are supported by NLP techniques and algorithms pre-trained in Python.

(1) Pre-processing the text acquired

The extracted texts are unstructured and need to be transformed into an understandable format before applying the word embedding algorithm. This step was performed with the support of Python’s NLTKFootnote 3 package, a suite of libraries for NLP. Words in the text were standardized with pre-established norms and were in accordance with the naming rules of WordNet. For example, since the same DE type can have different spellings (e.g. ‘testbed’ and ‘test bed’), they were pre-coded to be consistent to avoid information dilution due to spelling differences. Two tasks were performed in this step:

  1. 1.

    Tokenization The long text strings in each document are segmented into words. This task relies on the word tokenizer from NLTK.

  2. 2.

    Normalization This process transforms texts into a single canonical form and consists of a series of tasks such as (1) removing punctuations; (2) removing non-English words with the support of WordNetFootnote 4 which is a lexical database of English words; (3) removing general stop words like ‘the’ and ‘a’; and (4) recognising words’ grammatical form through part-of-speech (POS) tagging, then removing their inflectional endings and returning the base or dictionary format of the word accordingly through lemmatization (e.g. from ‘demonstrating’ to ‘demonstrate’) (Manning et al., 2008).

(2) Word embedding training and verification

Pre-processed texts were used to develop a language model based on Word2Vec with the support of Python Gensim package.Footnote 5 Word2Vec is a neural network model that converts each word in the corpus to a structured vector (Mikolov et al., 2013). It allows words with similar meanings to obtain a similar representation through a word embedding approach (Ni et al., 2022). By capturing subtle semantic relations between words, these models enable identifying concepts and relations from texts. For example, terms related to ‘testbed’ can be found by measuring cosine similarities between the vector ‘testbed’ and other vectors.

CBOW and SKIP-G are two main methods to implement word embeddings using Word2Vec (Mikolov et al., 2013). This study used the SKIP-G model (min count = 30, window size = 6). The objective is to identify candidate ontological components. Therefore, pairs of similar words determined by the model are the focus rather than achieving a high accuracy as it does not result in major differences in the similar words identified.

To verify the trained model, a simple set of keywords was derived for preliminary verification. These keywords were obtained by reviewing full articles during a structured literature review. To further evaluate the model performance in identifying similar keywords, these keywords were grouped such that they were expected to appear together when applying the similarity function:

\( S_{1}\) = [testbed, modular, heterogeneous, fidelity, federation]

\( S_{2}\) = [pilot-plant, pilot line, pilot-scale plant, pilot-facility, scale, demonstration, experiment]

\( S_{3}\) = [living lab, innovation lab, innovation, creative, stakeholder, user, collaboration, participation]

Through checking similar words of the listed words, potential keywords that are not included in the search string may appear, therefore requiring a search string refinement. In the experiment, two new terms, ‘sandbox’ and ‘testbench’ were identified and added to the search string (Table 2) for conducting a new search to enrich the text corpus with articles related to ‘sandbox’ and ‘testbench’. The new search was conducted on 7th July 2020, where 586 articles were added. The new text corpus consists of 4045 articles (total word count: 10,071,175; unique words: 36,523).

Ontology construction

(1) Knowledge extraction

With the language model trained, several algorithms were introduced to identify ontological components based on model outputs. These algorithms were adapted from similar works done by Mahmoud et al. (2018) and Albukhitan and Helmy (2016). This includes extracting domain words, identifying candidate concepts and properties from the domain words, building the ontology hierarchy and identifying concept-property relations. In this work, the BIRCH algorithm was used because it provides a scalable and efficient clustering algorithm for a large volume of text data.

Table 3 Seed terms used to extract domain words
Table 4 List of false-positive words
Algorithm 1
figure a

Algorithm for Domain Words Extraction.

\(^a\)findSimilarTerms(): find similar terms of the input seed term.

\(^b\)Similarity(): get cosine similarity value between two input terms.

  • Domain words extraction The first step is to extract domain words (Algorithm 1). Initial seed terms and the pre-trained language model are taken as inputs. Words related to the domain are extracted iteratively based on their relevance to the initial seed terms, represented as findSimilarTerms(seedModel) in the algorithm. In this work, we chose to measure the relevance between two words with the cosine similarity rather than the Euclidean distance since the magnitude of the word vectors (i.e., word counts) does matter, which follows the method as the previous studies, such as Kulmanov et al. (2020) and He et al. (2006). This was realized through the \(Word2Vec.most\_similar\) method in Python Gensim.Footnote 6 A similarity threshold value of 0.5 is employed to determine the inclusion of words that are deemed to be relevant to the domain. This value was set in order to obtain sufficient candidate terms initially. This threshold value can be adjusted according to the average and standard deviation of all the similar terms. Eventually, this algorithm returns a domain word list as the output for the next step. In this study, seed terms used (Table 3) were derived from the search string and full article reviews conducted in a structured literature review. Following Algorithm 1, 2256 unique words were identified to be relevant domain words.

  • Candidate concept and properties identification The second step is to identify two ontology components: concepts and properties from the domain word list. Algorithm 2 illustrates how domain words are organized to be concepts or properties. This step takes a POS tagger and the domain word list as the input. A POS tagger from the NLTK module will tag each word as a noun, verb, adjective, etc. Noun or adjective words are considered to be concepts, whereas verb words are taken to be properties. In this study, tagging the obtained words from the previous step gave 145 candidate properties and 2033 candidate concepts.

  • Ontology hierarchy building To build the ontology hierarchy, Balanced Iterative Reducing and Clustering (BIRCH) algorithmFootnote 7 (Zhang et al., 1996), which is a hierarchical clustering algorithm that exhibits better time performance for large datasets, was applied. BIRCH has two phases: (1) building cluster feature (CF) tree for initial clustering; and (2) global clustering phase that applies an existing clustering algorithm to further refine the initial clustering (Zhang et al., 1996). This step is supported by Python’s Scikit-learnFootnote 8 (version 0.23.1) module. In the hierarchy-building process, the vector representation for each identified concept was obtained from the language model and input to the BIRCH function for initial clustering (Algorithm 3). Agglomerative clusteringFootnote 9 was applied in the second phase for each initial cluster to obtain an ontology hierarchy. This clustering process generated nine CF-tree initially. Applying the agglomerative clustering algorithm produced different word clusters, which provides indications of concept groups because each group can relate to different themes (e.g., word clusters related to testbeds). We also plotted the dendrograms to obtain the word hierarchy that helps to determine the ontology structure.

Algorithm 2
figure b

Candidate Concept and Properties Identification.

\(^a\)wordTag(): tag the input term using a POS Tagger.

Algorithm 3
figure c

Ontology Hierarchy Building.

\(^a\)Model(): obtain the word vectors which will be used as inputs in BIRCH.

  • Concept-property relation identification Each concept has properties or attributes that describe it in an ontology. Therefore, the final step matches candidate properties with most related candidate concepts to provide inferences for generating axioms. This relation is determined by measuring the similarity between word vectors of concepts and properties from the pre-trained language model (Algorithm 4).

Algorithm 4
figure d

Concept-Property Relation Identification.

(2) Ontology population

Ontologies are often represented with Ontology Web Language (OWL) following the World Wide Web Consortium (W3C) standards (Lu et al., 2019; Montero Jiménez et al., 2021).In this study, OWL was chosen to model the DEO. Based on the hierarchical structure and concept-property relations obtained in the previous step, we manually populated the ontology in an editor. The ontology was coded using the open-source Protégé editor (version 5.5.0) (Porzel and Malaka, 2000). To respect the recommendation about reusing concepts from other ontologies, we used the Linked Open VocabularyFootnote 10 search engine and enriched the DEO with concepts from FOAF and SKOS. The full URI of the below-mentioned namespace prefixes can be seen in Table 5.

(3) Ontology validation

The constructed ontology aims to define and conceptualize DEs. Therefore, validation approaches that aim to assess the correctness and knowledge coverage of the developed ontology were deemed relevant and appropriate for this research. Therefore, both automated consistency checking and criteria-based validation approaches were selected due to their relevance and availability in this case. Task-based and data-driven approaches are not as relevant, moreover, this ontology was built from scratch since there is no benchmark ontology available. Similarly, expert validation is not applicable at the point when this research was conducted. These approaches were thus not considered.

The validation started with automated consistency checking. This is conducted to ensure that the logical axioms are consistent and there is no contradictory fact. To achieve this, the description logic reasoner HermiTFootnote 11 (version 1.4.3.456), a reasoning plugin in Protégé 5 was used. HermiT determines the consistency of the ontology and identifies subsumption relationships between classes (Shearer et al., 2008). The reasoner was used to debug the ontology along the ontology construction. The final results made by the reasoner revealed that no contradicting information was found in the DEO ontology. We also applied the OntOlogy Pitfall ScannerFootnote 12 during the ontology development, which provided great evidence of the design error-free of the DEO.

The criteria-based validation was also applied to evaluate a set of constructs iteratively during the ontology construction (Vrandečić 2009). Two criteria were excluded from this validation. First, the criteria of ontology adaptability was not assessed in this work as it was not relevant to the research question for which the DEO is developed. Second, external expert validation is needed to ensure the criteria of clarity. While our ontology includes well-annotated rdfs:commentFootnote 13 to effectively convey intended meanings, we acknowledge that external validation could offer greater clarity. We further detail this as a limitation in “Limitations and future research directions” section and explore opportunities for updates to the DEO. While we followed established approaches, external ontology validation may provide new avenues for future research and discovery.

The criteria-based validation of DEO is as follows:

  1. 1.

    Accuracy To ensure the descriptions in DEO agree with the domain knowledge, the DEO was cross-validated against several journal articles that discuss the conceptual ideas of DEs, as well as DE papers from different application domains. These articles were selected from SLR and conceptual articles with the highest citations in the domain. Validation results have provided evidence of correctness.

  2. 2.

    Completeness This research has successfully identified, and classified various named DEs and characterized them accordingly (“Demonstration environments ontology” section). It can thus be concluded that the completeness of DEO was achieved as the research objectives detailed in “Methodology” have been met.

  3. 3.

    Computation Efficiency The operation time of the HermiT reasoner was usually in seconds. This indicates that inferences can be done in a reasonable time and thus computation efficiency was achieved.

  4. 4.

    Conciseness No concepts in the same class were found greatly overlapping with each other (except for the equivalent class), so it was concluded that there were no redundant terms in the DEO and conciseness was met.

  5. 5.

    Consistency To ensure the logical axioms are satisfiable and consistent, it has to be possible to find a situation when all the axioms are true and there are no contradictions within the axioms. The output results of HermiT reasoner along the ontology development stages have verified this satisfaction and consistency.

To conclude, ontology validation has proven the validity of DEO from various aspects.

Ontology implementation, maintenance and application

New types of DEs can constantly appear in the literature or practices, creating the need for ontology maintenance. The ontology development is accomplished in GitHub as a public online repository. For traceability and transparency of the ontology development, the issue tracking system in GitHub is utilized as a communication channel for the maintenance as well as the future updates of the DEO by the public (Diehl et al., 2016; Turchet et al., 2020). This opens the opportunity for the community to validate and contribute to the DEO. The GitHub repository is freely available at https://purl.archive.org/demonstration-environments-ontology which is synced with our OSF project. For future use, this repository will be able to track the previous version of the DEO, while the latest version of DEO will always be accessible.

In addition, the application of the DEO is also demonstrated in this study with a real example through ontology instantiation, which is detailed in “Application of DEO” section.

Demonstration environments ontology

This section describes and highlights the building blocks of DEO that define and conceptualize DEs. Figure 2 provides an overview of the structure of the DEO. Figure 3 shows an example of the classes, object properties, axioms asserted and annotations of classes described in DEO.

Fig. 2
figure 2

Overview of the structure of Demonstration Environments Ontology (DEO)

Fig. 3
figure 3

Structure of DEO represented in the Protégé ontology editor software. (1) Class hierarchy (2) Annotation for class ’Testbed’ (3) Class relationships and axioms of ’Testbed’ (4) Object properties

Fig. 4
figure 4

DEO:DemonstrationEnvironment class structure

Classes

This ontology consists of five key classes:

  • DEO:DemonstrationEnvironment

  • DEO:Component

  • DEO:Characteristic

  • DEO:ApplicationDomain

  • DEO:ValuePartition

DEO:DemonstrationEnvironment is the class for named DEs. The four classes DEO:Component, DEO:Characteristic, DEO:ApplicationDomain and DEO:ValuePartition further define and characterize these named DEs.

Demonstration environment class

This class defines 15 named DEs as concepts, with 9 of them are unique DE types (Fig. 4). Those grouped in the grey circle are regarded as equivalent classes. For example, DEO:Pilot-scalePlant, DEO:PilotPlant and DEO:PilotLine are considered interchangeable. Similarly, DEO:InnovationLab has four other different names representing the same concept. Some named DEs can be further classified based on the utility: DEO:Sandbox is divided into DEO:GeneralSandbox and DEO:RegulatorySandbox; DEO:LivingLab is categorized into four subcategories based on who initiates the LL.

Component class

The component class is defined to describe the network surrounding different types of DEs and is further split into five major subclasses (Fig. 5, only expands this class up to 2nd degree of the subclass):

Fig. 5
figure 5

DEO:Component class structure (excerpt)

  1. 1.

    DEO:Objective This class defines a set of objectives that a DE application aims to fulfil. These objectives may include tasks like conducting assessments and diagnoses (DEO:Assessment &Diagnosis) to detect potentially suspicious behaviors.

  2. 2.

    foaf:Agent This class refers to actors that are involved in exploiting the DEs. For example, human beings or organizational institutions that design, develop, implement or utilize a DE. Here the well-known ontology FOAF (friend of a friend)Footnote 14 was reused according to Noy and McGuinness (2001). Once FOAF is connected to the DEO, it defines the foaf:Agent as things (e.g., person, group, software or physical artefact) that do stuff. Two of its subclasses foaf:Person and foaf:Organization were also included to further specify the type of agents that initiate the use, support or management of a DE.

  3. 3.

    DEO:Environment This class represents a specific DE that is working in. Based on the degree of realism, the environment is classified into (1) DEO:ModeledContext which usually refers to the laboratory or simulated program; (2) DEO:Real-lifeContext such as the campus or the city which is consistent with the final application environment; and (3) DEO:Semi-realisticContext which lies in between the formers (e.g. a partially simulated environment placed in a real-life context like the scaled-down plant).

  4. 4.

    DEO:Outcome This class defines a set of tangible outcomes such as a product or a prototype, or intangible outcomes such as a service or a concept of the demonstration.

  5. 5.

    DEO:Tool &Facility This class pertains to the tools and facilities employed by a DE to facilitate its operations and processes. These tools and facilities might encompass software applications or machinery.

Application domain class

The DEO:ApplicationDomain includes a set of application domains of DEs that are most commonly identified in the text corpus. For example, living labs are commonly applied in the domain of public affairs. The domains included are not exhaustive but representative to describe the most common areas of DE applications. This class includes defined instances such as Engineering and Biochemistry.

Fig. 6
figure 6

Using the MoteLab example to instantiate the DEO

Characteristic class

This class includes a set of terms that are used to describe the general characteristics that are shared within a DE class. These characteristics can belong to subclass DEO:ProcessCharacteristic that characterizes the process of applying DE, or subclass DEO:StateCharacteristic which describes the inherent features of DEs, or subclass DEO:RequirementCharacteristic which includes a set of characteristics that need to be possessed by a DE. In addition, an instance of a characteristic can belong to multiple subclasses.

Value partition class

A value partition refers to a pattern that represents the value properties of DEs. This class defines partitions for three values: DEO:SizeScale, DEO:Lifespan and DEO:TargetMaturity of the DEs. DEO:SizeScale class is used to describe the size of the DEs. This class is further split into two subclasses: DEO:Large-scale and DEO:Small-scale. The former one includes scales (defined as instances) such as DEO:Production-scale and DEO:Demonstration-Scale, which are often used to represent DEs that are applied in the last stage of demonstration. In addition, DEO:Lifespan describes the period during which a DE is functional, and DEO:TargetMaturity indicates if a DE deals with rather mature (market-ready) or nascent technologies.

Application of DEO

This section aims to illustrate the applicability of DEO by first presenting a real example for ontology instantiation, and then demonstrating how SPARQLFootnote 15 can be used to search defined knowledge in the DEO.

Fig. 7
figure 7

(Left) A screenshot of instance ’Malware_analysis’ from Protégé (1) definition of ‘Malware_analysis’, (2) concept that ‘Malware_analysis’ belongs to, (3) asserted object relations between ‘Malware_analysis’ and DEO: DemonstrationEnvironment instances; (Right) Interface of corresponding SPARQL query, (4) query, and (5) query results display

The MoteLab: a wireless sensor network testbed example

To test the performance of the DEO and better present its structure, a real example of the testbed called MoteLab - a wireless sensor network testbed developed by Werner-Allen et al. (2005) is presented. As shown in Fig. 6, the graph firstly introduces DEO:MoteLab as a specific example of a testbed. By exploiting the DEO, a semantic network to describe and characterize this DE is set up.

  1. 1.

    Application domains and the value measurement Wireless sensor network has emerged to be a new area of research in the field of Computer Science. In this case, the MoteLab is used in the application domain of DEO:ComputerScience. MoteLab has been largely deployed at a building in Harvard University, so its size scale is DEO:Large-scale. Moreover, Werner-Allen et al. (2005) mention the next phase of this testbed, as such, the maturity of the MoteLab is still DEO:Semi-mature.

  2. 2.

    Components involved in the MoteLab Researchers, software developers and organizations such as Intel Research and Havard University were involved in the life-cycle of MoteLab development. All these instances are pointed to DEO:MoteLab through the DEO:hasActor predicate. Moreover, the FOAF ontology then provides the human name values of foaf:Person instances through foaf:name. A set of hardware devices and software components such as PHP-coded web server and DBLogger are defined as instances of DEO:Hardware &Software that are linked to DEO:MoteLab through DEO:hasTool relationship in the DEO. DEO:hasObjective is connected with DEO:Experimentation and DEO:Teaching to show the objectives of developing this testbed. The corresponding outcome is hence a DEO:WirelessSensorNetwork, it is defined to be an instance of the class DEO:System which is a subclass of DEO:TangibleOutcome. Besides the MoteLab has been physically deployed in DEO:MaxwellDworkingBuilding which is a real-life environment.

  3. 3.

    Characteristics of the MoteLab DEO:hasCharacteristic links with several adjectives to describe the characteristics of MoteLab. As described in the original article (Werner-Allen et al., 2005), the MoteLab was designed in a way that enables users’ real-time access through several manners to perform real-time data analysis. As such, DEO:Real-time is defined to be an instance of the DEO:ProcessCharacteristic class. The article also mentioned the requirement when designing the MoteLab: the testbed has to be designed for realistic wireless sensor networks. Therefore, we instantiate DEO:Realistic within the DEO:RequirementCharacteristic. Besides the MoteLab has been designed to be open-source, automatically programmable and it is a physical thing that involves several devices. Open-source, Programmable and Physical are thus instances of DEO:StateCharacteristic that describe the states of the MoteLab.

SPARQL-based knowledge searching

This section explains how DEO can be used for knowledge searching through querying with SPARQL, which is a query language that supports the retrieval and manipulation of data stored in RDFFootnote 16 format. The DEO has established a conceptual framework for DEs. Through proper instantiation of classes and concepts, the DEO can be further used as a dictionary to search for defined knowledge given search strings coded in SPARQL syntax. To illustrate, this part presents a use case of searching for specific DEs that support the objective of assessment and diagnosis. This use case is selected because one typical user story could be a practitioner looking for suitable DEs to satisfy the innovation objective.

The full user story is described as follows: An IT maintenance staff is looking for DEs that support malware analysis. Figure 7 (top) shows the instantiation of DEO for the given use case: Malware_analysis is an instance (specific task) of DEO:Assessment &Diagnosis. Several real examples (instances) of DEs that can meet this objective have been created and corresponding object properties have been defined (For example, App_testbench which is an instance of DEO:Testbench) (Liang et al., 2014; Ahmad et al., 2016; Caranica et al., 2019).

Protégé provides the SPARQL Query tab as a development environment for creating SPARQL queries with the built ontology (Fig. 7 bottom). This use case presents searching instances of DEO:DemonstrationEnvironment that provide assessment and diagnosis of products, and look for which specific task they can perform and what DE type they are of.

It has been demonstrated how the DEO can be instantiated and used as a dictionary for querying defined knowledge. The next step of this user story could be searching for information on any of the DEs in the query result. This can be done by a similar instantiation step shown in “The MoteLab: a wireless sensor network testbed example” section and then applying SPARQL again to search the information.

Discussion

A multi-dimensional framework for conceptualising DEs was developed, with nine types of DEs being differentiated based on the four aspects which are domains to be applied, the network components involved, their characteristics and the value attributes they possess. The DEO provides a representative conceptualization of DEs with certain statistical significance. DEO has been designed and developed in a way such that it meets the four objectives as stated in “Methodology” section:

  1. 1.

    Providing a structure to define and differentiate DEs by determining aspects that characterize different DEs The structure of DEO, where the class of DE is connected with four other classes (Fig. 2), provides a standardized framework for defining DEs of different categories. In this regard, a DE can be described and characterized by its components, its characteristics, the contexts it is being applied, and dimensional values like its size. Such a standardized framework enables DEs of different types to be viewed and compared together, spurring future research to have integrated views of different DE types rather than focusing on a DE solely.

  2. 2.

    Ensuring the semantics are clearly defined such that there is no ambiguity in terminology and the meanings in the knowledge representation are clear; redundancy in relevant terminologies have to be captured by this ontology As described in Ontology Validation (“Ontology construction” section), the clarity of DEO has been examined to ensure that there is no ambiguity in defined vocabularies and relations. In addition, ontology has captured and documented redundant terminologies that have been used to describe the same thing. Differences in spelling have also been annotated if exist. This includes ‘living lab’, ‘living laboratory’ and ‘living-lab’ which all represent the same demonstration environment. Differences in wording have been captured as well. For example, ‘pilot plant’ and ‘pilot-scale plant’ are indeed the same with a certain statistical significance level. This is said because the language model trained in this study has indicated that these two terminologies can be considered equivalent given the text corpus we have. Annotating these terminologies enables the heterogeneity in terminologies to be reduced in future DE studies, or at least this heterogeneity could be traced.

  3. 3.

    Offering means for searching and querying information, and supporting activities such as searching for and selecting suitable DEs, and proposing feasible DE designsSPARQL-based knowledge searching” section presents a use case of how DEO can be employed for knowledge searching using SPARQL. This use case demonstrates the applicability of DEO to be used for querying and searching defined information. The capability of performing querying enables DEO to be deployed as a tool for practitioners like decision-makers to identify suitable DEs given certain conditions, for example, the innovation goals they seek to achieve. Practitioners like LL facilitators can also use insights gained from querying the ontology to design and prepare the DEs. For example, the developed ontology specifies the network components surrounding the DEs, such as who is involved and what tools to use. Instantiating the DEO with real-life LL examples would help to satisfy this need.

  4. 4.

    Allowing the semantic model to be understood and interpreted by different users, applications and systems Since DEO is built upon formal OWL-based specifications, it ensures the information is interoperable across a wide variety of tools and platforms like Protégé by different users.

In addition to meeting the objectives, it is noteworthy that even though the DEO has defined nine categories of DEs, the classification and characterization presented should be considered as general-typical rather than absolute. There is no absolute independence when applying different DEs. For example, the typical user-centred and real-life context features from the living labs can be incorporated into a testbed for large-scale demonstration, one instance is the smart city testbed for carrying out a controlled IoT experimentation at the city level (Sanchez et al., 2014). Further instantiations of the DEO would enable it to capture both the generality and granularity of the DEs at an instance level.

Conclusion

The evidence for the need for a standardized framework for defining and characterizing DEs is compelling due to the two gaps outlined above, namely (1) the heterogeneity of DE domain knowledge, and (2) the lack of a holistic view of different types of DEs. Therefore, this paper offers a systematic and quantitative method of realizing such a framework. This paper describes the DEO in OWL for defining and characterizing nine different DE types. The creation of this ontology is motivated by the heterogeneous terminologies and inconsistent interpretations of different DEs in existing studies. Instead of relying on a limited number of experts, this study attempts to incorporate as much diverse knowledge as possible to expand the pool of collective intelligence by applying word embedding on relevant peer-reviewed journal articles to construct the ontology. Moreover, the structure and content of the DEO have been evaluated based on automatic consistency checking, criteria-based validation and cross-validation with a set of relevant journal articles. The result of ontology validation indicates that the DEO is consistent, effective and follows good practice.

Contributions

This study makes two key contributions to the innovation management literature. First, the DEO provides standardization of terminologies related to DEs, which solves heterogeneity in the description of various names of DEs across domains. Second, the DEO also offers a common understanding of DEs through a systematic framework, which enables different DEs to be benchmarked together with the established dimensions: domains to be applied, the network components involved, their characteristics and the value attributes they possess.

This study also has practical implications for innovation managers and practitioners. Defining and characterizing DEs is essential for practitioners and researchers working in different fields, as it provides a common understanding of the essential features of DEs. This shared understanding enables effective benchmarking of different DEs across established dimensions, which is crucial for assessing the effectiveness of new technologies, products, or services in a consistent and reliable manner.

Moreover, distinguishing between living labs, pilot plants, and testbeds and characterizing these DEs is vital for ensuring that practitioners use the right type of testing environment for the specific technology, product, or service being developed. This is critical for increasing the accuracy and reliability of testing results and ensuring that the final product or service meets the required standards and satisfies the needs of the end users. By using the appropriate DE, practitioners can test new technologies, products, or services in a realistic environment, with real users, in real-life scenarios (in the case of living labs), or in a controlled environment (in the case of pilot plants and testbeds).

Ultimately, a better understanding of DEs and the appropriate use of different types of testing environments can help to accelerate the development and deployment of new technologies, products, and services, and promote innovation in various fields.

Limitations and future research directions

This study has some limitations and thus we could point out potential avenues for the development of future research. First, although we collected the data from peer-reviewed journal articles to ensure the data diversity, quality and statistical significance, some subjective judgements from domain experts can help in refining and validating the DEO further. Nevertheless, our ontology is available in the GitHub repository, enabling open contributions from the community through pull requests to suggest contributions to the DEO. Second, the DEO has been constructed in a way such that it entirely focuses on DEs and only includes necessary innovation-related classes (e.g., DEO:Outcome, DEO:Tool &Facility) to define and characterize the DEs (DEO:DemonstrationEnvironments). The DEO can be further expanded to cover more components in innovation management through integrating with other innovation-related ontologies, such as those reviewed in “inlinkInnovation-related ontologiesInnovationspsrelatedOntologies” section. Third, the Word2Vec model used in this study may fail to capture creative and insightful DE knowledge of the few as Schnabel et al. (2015) find that the word frequency information is contained in the word embeddings. A more sophisticated algorithm for improving the word embedding model performance is needed. Future research can be aimed at exploring algorithms for improving the word embedding model performance with low-frequency words. Finally, since the DEO first provides a standardized framework with formal and semantic representation for DEs, it serves to be an important foundation for exploring DEs in the future. Future work direction should be directed to disseminate the DEO to the industry and academia for testing its performance with real-life use cases and inputs to further improve the model usability and enrich the contents in the field of DEs.