1 Introduction

The KGs are increasingly used in research and business, while their development and deployment present a close association with Semantic Web (SW) technologies (including Ontologies and Linked (Open) Data), large-scale data analytics, and cloud computing [1]. As mentioned in Ehrlinger and Wöß [1], KGs have been in the focus of research since the introduction of Google’s Knowledge Graph in 2012, resulting in a variety of descriptions and definitions of the term such as the ones provided by Paulheim [2] and Ehrlinger and Wöß [1]. The description of Kejriwal [3] states that the requirements a graph must fill to be considered a KG, are (i) its meaning to be expressed as structure, (ii) its statements to be unambiguous, and (iii) to use a limited set of relation types.

The technologies that can be deployed in order to build a KG, include (i) knowledge representation and reasoning (languages, schema, and standard vocabularies), (ii) knowledge storage (graph databases and repositories), (iii) knowledge engineering (methodologies, editors, and design patterns), and (iv) knowledge learning (including schema learning and population) [4]. Different platforms and suites, which partially or fully support the aforementioned technologies, have been developed, thus providing the necessary tools and processes for the development, maintenance, and use of KGs (e.g., Neo4j suite [5], OpenLink Virtuoso platform [6], RDFox [7]).

The development of KGs presents a close association with ontologies. In most cases (DBpedia [8], Wikidata [9], YAGO [10], Google KG [11]) ontologies constitute the backbone of KGs, since they are the semantic model or models that KGs incorporate. Formal ontologies are currently the most popular technology for developing semantic models to represent a particular domain of knowledge formally and explicitly.

Several methodologies have been proposed through the years for engineering ontologies [12,13,14,15,16,17,18,19,20,21,22]. The ontology lifecycle, which includes feasibility analysis, identification of goals, requirements specification, implementation, evaluation, and maintenance, is defined almost similarly by different ontology engineering methodologies (OEMs), and it is partially or fully supported by ontology engineering (OE) tools [19]. Based on the obvious association of ontologies and KGs, Carriero et al. [23] suggested that the ontology and the related KG can be both developed following the engineering principles, or similar/analogous tasks and steps, of the same OEM. This approach sets the methodological steps of the KG lifecycle, including its semantic model, in a way that is consistent and—probably—familiar to the specialists who are involved in the development of a KG. According to several methodological approaches, OE is mainly driven by the ontology engineer, i.e., the person who has the knowledge/expertise to define ontological specifications and to coordinate an OE task. However, the role and active involvement in the ontology lifecycle of (i) domain experts, who have the knowledge/expertise of the domain and/or data sources, as well as (ii) knowledge workers, who exploit the ontology in ‘operational’ conditions (e.g., solve problems, perform data-driven analysis tasks) is considered equally useful and essential for human-centered and collaborative approaches [13, 15, 19, 22,23,24].

Collaborative OEMs define, in a systematic way, phases, tasks, and workflows which emphasize the active and decisive involvement of ontology engineers, knowledge workers, and domain experts throughout OE process via their close and continuous collaboration [19]. This approach significantly empowers the knowledge workers and domain experts, people who have the knowledge and the expertise of the domain of interest, though they may be not familiar with the (i) formal representation languages, (ii) knowledge engineering principles, and iii) methods for constructing and synthesizing ontologies [15, 19, 22, 24]. Knowledge workers and domain experts actively participate in the collaborative OE processes, along with ontology engineers, and they are able to develop, evaluate, and evolve ontologies individually and conversationally with their co-workers, according to their skills, knowledge base, and context of work [19, 22].

Since collaborative OEMs are considered beneficial for the complete and consistent development of ontologies in a human-centered manner [15, 19], their principles and tasks could also be adapted for the development of KGs. Thus, the collaborative approach could involve participants of different levels of expertise related to KG development, enriching, and maintenance, in a continuous and systematic collaboration.

This paper presents the methodology of Collaborative and Hybrid Engineering of Knowledge Graphs (CHEKG—pronounced “check”) which was first introduced as ongoing work in Moraitou et al. [25]. CHEKG constitutes a hybrid (schema-center/top-down and data-driven/bottom-up), collaborative, and iterative approach for KG development. The CHEKG methodology is based on the main methodological phases of the latest version of ext-HCOME OE methodology [26], while it adjusts and expands the individual processes and tasks of each phase according to the KG development specialized requirements (modularity of KG’s model and content, agility, data quality, and bias). Although there are several other agile OEMs such as UPONLite [21], SAMOD [20], AGIScont [22], and XD [27] that could be exploited or adapted for KG development, the presented work is novel since, to the best of our knowledge, another effort to adapt a collaborative, iterative, and agile methodology to propose a methodology for engineering domain-specific modular and fair KGs, has not been presented elsewhere yet. Apart from the presentation of the methodology per se, the paper presents current work regarding the deployment of CHEKG methodology for the development of domain-specific KGs for the representation of semantic trajectories generated from UAVs data. This work is motivated by, and evaluated based on, the application domain of UAV missions for documenting regions of cultural heritage interest, which is presented in Kotis et al. [28].

The structure of the paper is as follows: Sect. 2 reviews existing methodological steps for KG development, while it discusses the lack of human-centered and specialized tasks for KG maintenance. Section 3 extensively presents phases, processes, and tasks of CHEKG methodology. Section 4 describes the main results of the implementation and evaluation of the CHEKG methodology with domain-specific KG development; Sect. 5 discusses the findings and limitations of the proposed approach. Finally, Sect. 6 concludes the paper.

2 Related work

In addition to the ext-HCOME, other collaborative and agile OEMs support ontology lifecycle in a systematic way, which emphasize the active and decisive involvement of ontology engineers, knowledge workers, and domain experts throughout OE process via their close and continuous collaboration [19]. Their structure comprises three main phases: (a) ontology specification, (b) ontology development, and (c) ontology exploitation and evaluation phase. Common tasks include the definition of ontology’s scope and aim, ontological definitions’ reuse, model and instance definition, validation through exploitation in use cases, and refinement/maintenance through iteration. The following representative OEMs are briefly presented as example related work, selected mainly for being recently proposed methodologies, that incorporate agile principles and underline the importance of collaboration by providing clear instructions and reducing the dependence on ontology engineers.

UPON Lite is an OEM emphasizing a participative social approach. It reduces the role of ontology engineers, allowing domain experts, knowledge workers, and end-users to collaboratively build ontologies using user-friendly tools. The six-step process includes identifying terminology, defining a glossary, generating a concept taxonomy, connecting entities, defining parthood, and developing the ontology. The methodology supports agile collaboration, exploitation, and evaluation of ontologies.

SAMOD proposes a simplified and agile approach to ontology development, inspired by test-driven development processes. The methodology is iterative, focusing on documented ontology creation from typical examples of domain descriptions, using motivating scenarios and competency questions (CQs). Collaboration occurs between domain experts and ontology engineers at the initial steps. SAMOD employs an evolving prototype approach, where ontologists collect requirements from domain experts, develop an initial model, and iteratively refine it based on scenarios until it satisfies all CQs.

AgiSCOnt goes a few steps further, proposing an end-to-end OE process, addressing project objectives, tools, scheduling, budgeting, and resource allocation. The three steps of AgiSCOnt involve analysis and conceptualization, development and testing, and ontology use and updating. The methodology encourages collaboration between ontology engineers and domain experts, leveraging knowledge elicitation techniques, conceptual maps, and CQs to develop ontologies iteratively, with a focus on the reuse of and alignment to existing models. To the best of our knowledge, this is the latest new and maintained collaborative and agile OEM (along with ext-HCOME).

KGs are exploited in order to semantically enrich large amounts of data found in various data silos, adding value to it, so that it can be (re)used in a meaningful, machine-processable, and more intelligent way [29]. The processes/tasks for the development and maintenance of a KG may vary, and therefore different guidelines and extensive methodologies have already been proposed. The following paragraphs present related work that is selected due to the fact that they are comprehensive methodologies covering the entire KG engineering lifecycle. More specifically, they were selected based on their utility and reproducibility across a variety of scenarios, while they are not applicable only to a specific domain. as others ([30,31,32,33]) or based solely on data-driven approaches, as others ([34,35,36]). Our literature review focused on the latest five-year period, within various sources, including Google and Semantic Scholar, IEEE Explore, ACM Library, ScienceDirect, and Scopus. The sources were queried with the keywords “Knowledge Graph” followed by “development,” “construction,” “engineering,” “lifecycle,” and “methodology.”

As Fensel et al. [29] suggest, the major steps of an overall process model for KG engineering are (i) knowledge creation, which is a knowledge acquisition phase that establishes the core data for a KG, (ii) knowledge hosting, (iii) knowledge curation, (iv) knowledge deployment, which is the actual application of the KG in a specific application domain for problem-solving.

A work that discusses the need for guidance on KG development, as they are widely used in various AI-driven tasks with large-scale data, is presented in Tamašauskaitė and Growth [37]. It aims to provide guidance in planning and managing the process of KG development, by synthesizing common steps described in academic literature and presenting a KG development and maintenance process. The process involves steps that include data identification, ontology creation, data mapping, knowledge extraction, visualization, KG refinement, and the deployment and maintenance of the KG.

Apart from the suggestion of particular methodological steps for KG development, a few extended methodological approaches have also been suggested. For instance, a recent work presents a bottom-up approach to curate entity-relation pairs and construct KGs and question-answering models for cybersecurity education [38]. The methodology includes three main phases: (i) knowledge acquisition, (ii) knowledge storage, and (iii) knowledge consumption.

Regarding the use of fully and well-defined methodologies for KG development, the exploitation of existing OEMs has been suggested. Particularly, the Extreme Design (XD) methodology has been used for the development of ArCo KG and its underlying ontology [23]. The methodology includes a set of major procedures: (i) requirement engineering, (ii) CQs, (iii) matching CQs to ontology design patterns (ODPs), (iv) testing and integration, (v) evaluation. Sequeda and Lassila [39] describe the phases of Designing and Building Enterprise KG and identify the involved people, the KG management, and the necessary tools for developing a KG.

The different guidelines and methodologies present some similarities, especially regarding the main tasks/processes that must be followed for KG development. For instance, the identification of KG requirements, the definition of data that it will capture, the efficient choice or development of the underlying knowledge model, the consistent evaluation and correction of the KG, the enrichment and augmentation of the KG, and finally the implementation of the KG for specific services or processes, are all crucial/important tasks. Additionally, the data and knowledge (conjointly or individually) are at the center of the development since they are vital parts of the KG.

Although bottom-up (data-driven) and top-down (schema-centric) approaches for the conceptualization of the knowledge that the KG captures are both necessary (constituting a hybrid knowledge conceptualization approach), the role and specific activities that involved people must follow have not been described in detail and have not been emphasized in methodological phases and steps. Emphasizing a human-centered, collaborative, and hybrid approach that focuses on specific domains, could empower the involved stakeholders, namely domain experts, knowledge engineers, knowledge workers, and bias/fairness experts, to be continuously involved in the KG engineering lifecycle. Such an approach could incorporate all the different KG development tasks and organize them in unambiguous phases, clarifying the roles of the involved members of the development team, exploiting their specialized knowledge in conceptualization, data/knowledge acquisition, KG deployment, and KG evaluation.

A comparison of the aforementioned related work and their mapping to the proposed methodology is presented in the discussion section.

3 The CHEKG methodology

The HCOME constitutes a human-centered collaborative OEM, according to which the ontologies are engineered individually, as well as collaboratively. The HCOME supports the involvement of knowledge workers, and it requires the use of tools which facilitate the management and interaction with conceptualizations in a direct and iterative way. The methodology is organized into three main phases, namely, specification, conceptualization, and exploitation/evaluation, emphasizing on discussion and argumentation of the participants over the conceptualization, the detailed versioning of the specifications, and the overall management of the developed ontology. The basic tasks of HCOME are enriched by data-driven (bottom-up) conceptualization tasks [40], supported by the learning of seed ontologies. A stand-alone OE integrated environment, namely HCONE, has been used until 2010 to support management and versioning tasks in the individual space of participants, while the Semantic Media Wiki-based shared environment, namely Shared-HCONE [41], has been used to support evaluation tasks and the argumentation-based discussions in the collaborative space of the participants. Today, the HCOME is supported by alternative tools, such as Protégé [42], Web-Protégé, email lists, and shared cloud workspaces, mainly for the collaborative space tasks [19]. The latest version of the ext-HCOME methodology [26] has been updated with the modularization and bias-assessment/mitigation tasks.

Based on the HCOME methodology, a hybrid (schema-centric/top-down and data-driven/bottom-up), human-centered, collaborative, iterative, and agile methodology for the engineering of modular and fair KGs contributes to all phases of the KG engineering lifecycle: from the specification of a KG to its creation, exploitation, and evaluation. The proposed methodology provides a distinction between specific processes of each phase, while it breaks down the processes into tasks that can be performed either individually (in a personal engineering space) or collaboratively by all members of the KG engineering team. It is considered that team members may engineer KGs in their personal space, while they perform individual tasks, but they may also exploit a shared engineering space in cases where they perform collaborative/argumentation tasks. The tasks that can be performed in shared spaces are tasks which (i) can—technically—be performed using a collaborative file/software and (ii) depend on synchronous and asynchronous discussion and contribution by different members. Collaboration (and decentralized workflow) could be supported by specific collaborative tools such as WebProtege, git repositories such as GitHub, and cloud collaboration workspaces such as Google workspace, as well as by using emails and videoconferencing. It is possible for a team member to initiate any task either in a personal or a shared space or take part in any other task which has already been initiated by other members. Shared space is indicated with the letter ‘S’ and personal space is indicated with the letter ‘P’ in the description of each task.

Within the process of KG engineering, there are both mandatory and optional tasks, each serving distinct purposes. Mandatory tasks such as data modeling or storing and querying knowledge are essential for the development and usage of every KG, without which a KG could not exist or be considered a KG. Optional tasks, on the other hand, such as the semantic enrichment and the assessment/mitigation of bias, may be used to refine and add value to a KG, can be deferred to subsequent iterations or can be applied according to the context of work. The optionality of tasks’ execution is mainly determined by a combination of project-specific factors, including resources, requirements, constraints, and domain complexity. Mandatory tasks must be performed in the specific order that are mentioned for the first iteration. However, since CHEKG follows an iterative approach, additional iterations may start from any task that is required. Mandatory tasks are indicated with the letter ‘M’ and optional with the letter ‘O’ in the description of each task. The following sections describe the processes and tasks of each phase of CHEKG methodology, as depicted in Figs. 1 and 2.

Fig. 1
figure 1

Phases and processes of the CHEKG methodology

Fig. 2
figure 2

Tasks of CHEKG methodology

3.1 KG specification phase

The KG specification phase establishes the involved team, as well as the context of work regarding the KG development. During this phase, the members of the team are identified and their role in the whole endeavor is defined. Consequently, the involved team starts discussions over the aim, scope, and requirements of the KG, while it composes specification documents, for example the Ontology Requirements Specification Document [43], that describe the aforementioned—agreed—information. Additionally, during this phase, the main data sources that will be exploited for KG development are detected. This phase may start from a member of the team or a small-core group of the team (e.g., the knowledge engineers) who has made some preliminary work for the identification of the KG model and data, and who needs the contribution of other colleagues and domain experts for the validation and elaboration of this work. The KG specification phase (Phase 1) is mainly performed within the shared space, and it includes:

Process 1.1. The specification of the involved team. This process is further analyzed in the following tasks:

  • Task 1.1.1: Identify collaborators (i.e., domain experts, knowledge engineers, knowledge workers, bias experts), in order to determine the people who will be involved and their background and interest in the endeavor. (S, M)

  • Task 1.1.2: Determine team’s composition and members’ roles to establish the work team and organize (if needed) subgroups of the team which could contribute to different tasks or have a different level of involvement in different tasks. (S, M)

Process 1.2. The specification of aim, scope, and requirements of the KG. This process is further analyzed in the following tasks:

  • Task 1.2.1: Specify the aim and scope of required and/or available data, which is an essential task in order to (i) establish a common perception of the domain that the KG will cover and (ii) agree upon the reason for the KG creation, between the different members of the team. (S, M)

  • Task 1.2.2: Identify the main data sources, such as datasets, taxonomies, and other information, which will be the initial sources that will supply the KG with data. The sources may be proprietary, open, or commercially available. They should be chosen taking into account the specified aim and scope of the KG, as well as the considerations of future KG maintenance (e.g., possibility of KG update with new data from the available data sources). The sources are also used during KG model establishment and evaluation. (S, M)

  • Task 1.2.3: Discuss and specify the design requirements of the KG that will be commonly understood and accepted by the work team. Important design requirements that need to be specified in the design of the KG are: Scalability, Performance, Interoperability, Semantic Expressiveness, Reusability, Data Quality, Privacy. (S, M)

  • Task 1.2.4: Establish domain-related questions to be answered by exploiting the engineered KG (eventually) and formulate CQs. The CQs will be useful for the KG model development and KG evaluation. This task is highly recommended, and it is considered best practice. It is set as optional since in simplistic, well-defined, or experimental cases it can be loosely covered by tasks 1.2.1 and 1.2.3. (S, O)

  • Task 1.2.5: Produce specification documents for the KG, in order to record and share agreed specifications in appropriate collaborative forms and documents (e.g., shared cloud workspaces). (S & P, M)

It is worth mentioning that task 1.2.5 can be performed either in a shared or a personal space, but in any case, it must be communicated and agreed upon by all the members of the involved team.

The possible roles of members of the KG engineering team could be: Domain Expert, Knowledge engineer/Ontologist, Data Scientist, Software Engineer, Quality Assurance Specialist, Bias Expert, Privacy Expert, Security Engineer, and End-User/Customer. Roles may vary depending on the project and the involved team, while it is possible that individuals have overlapping roles/responsibilities or change roles during the engineering process. For example, a Domain Expert could also be assigned the role of an End-User. As it is also proposed in HCOME methodology, in CHEKG methodology the roles of all stakeholders are equal, and all participants must be involved toward a true collaborative engineering experience.

3.2 KG development phase

The KG development phase follows the KG specification phase, during which the involved team, either working on a shared or a personal space, develops the model, the data, and the infrastructure which will store the KG. It is possible that different members or subgroups of the involved team may focus on one or more areas of work, e.g., ontology engineers may focus on explicit knowledge creation of the KG (i.e., the semantic model of the KG, in other words, the KG’s schema). The KG development phase (Phase 2) includes:

Process 2.1. The creation of explicit knowledge, which refers to the semantic model that the KG incorporates. This process is analyzed in the following tasks:

  • Task 2.1.1: Consult experts by discussion, in order to better understand the domain of interest. Identify concepts that can be grouped as modules of the subdomain conceptualization, based on the project requirements and objectives. This is a human-centric approach for the KG model development. This task could be omitted in simplistic cases or in cases of reusing a well-established schema that covers the domain of interest (S, O).

  • Task 2.1.2: Gather, analyze, and clean data of the identified main data sources, in order to identify central concepts and relations, as well as to outline the content that the KG model should represent. The cleaning and correction of data will improve their quality, making them most applicable for the analysis and later KG instantiation. The cleaning and correction may include removing invalid or meaningless entries, adjusting data fields to accommodate multiple values, fixing inconsistencies, etc. This is a data-driven approach for the KG model development (P, M).

  • Task 2.1.3: Learn a kick-off semantic model exploiting algorithms over the pre-processed data. This is a data-driven approach for the KG model development. Although this task can be helpful in the schema definition process, it is not critical, and its results can be covered by tasks 2.1.4, 2.1.5, 2.1.7 (P, O).

  • Task 2.1.4: Reuse semantic models that may be either relevant to the domain of interest or used/embedded by the identified sources. This task could include the analysis of different data schemata of the sources and the import of ontologies, taxonomies, thesauri, etc. (either parts of them or as a whole). The discovery of different semantic models relevant to the domain of interest using libraries of models or ontology design patterns may be performed in a systematic way, e.g., searching using key-terms and concepts that have been identified during initial stages of work or during preliminary discussions with experts. This is a schema-centric approach for the KG model development. This task is fundamental for the semantic model development process, and it should be performed if possible. However, it is considered optional in simplistic scenarios or cases where the domain of interest has not undergone in-depth study, or when the existing semantic models lack the necessary degree of semantic expressivity (S & P, O).

  • Task 2.1.5: Consult generic top-level semantic models (e.g., DOLCE, WordNet, DBpedia ontology), in order to better understand formal semantics of the domain of interest. This is a schema-centric approach for the KG model development. This task is considered optional as it could be covered by task 1.2.4 (P, O).

  • Task 2.1.6: Consult the kick-off semantic model, in order to (i) identify key-terms and concepts for the model and (ii) enrich the model under development. This task is optional since it depends on tasks 1.2.3 (not critical) (P, O).

  • Task 2.1.7: Specify and formalize the semantic model of the KG, in order to have a formal representation of the KG conceptualization. A part of this process is to develop the modules and seamlessly interlink them within the semantic model to ensure a comprehensive representation of the domain of interest. This may include either the engineering of the model from scratch or reusing and engineering the imported semantic models, top-level semantic models, and kick-off semantic models (P, M).

  • Task 2.1.8: Merge and compare different versions of the semantic model of the KG, to support its reuse and evolution. Especially in cases where the participants work in personal and shared space for the development of the model, it is very important to compare and merge the different versions that they have produced. This is set as optional, since it could be performed in subsequent iterations of the process or not (not critical) (P, O).

  • Task 2.1.9: Add documentation to the semantic model of the KG, with further comments, examples, and specification details which would make it more understandable to other people. This task is recommended, but it is set as optional since it could be performed in subsequent iterations of the process or not (not critical) (P, O).

  • Task 2.1.10: Discuss the specified semantic model with domain experts, in order to verify the modeling and designing choices and spot gaps and redundancy (S, M).

The process of developing the semantic model is already integrated in CHEKG methodology; thus, there is no need to reuse an external OEM for this process. This integrated process is based on ext-HCOME OEM.

Process 2.2. Create instance data of the KG. This process is further analyzed in the following tasks:

  • Task 2.2.1: Create instance data (KG’s data) via the semantic annotation of identified sources, using the produced semantic model of the KG (e.g., RDFization of sources). Particularly, CHEKG refers to the two main approaches to populate the KG with instance data: (i) mapping structured, relational data (e.g., CSV files, databases) to the semantic model of the KG using mapping languages/methods (e.g., RML, R2RML, SPARQLgenerate [44], RDF-Gen [45]) and (ii) extract data from unstructured sources (e.g., text files) in an automatic manner using machine learning algorithms and the semantic model of the KG (P, M).

  • Task 2.2.2: Validate produced data, in order to identify any modeling mistakes (e.g., using RDF shapes with SHACL) (S & P, M).

  • Task 2.2.3: Integrate data that are provided by different main data sources in order to be represented with the semantic model of the KG. This task is mandatory in cases where more than one data sources have been recognized and incorporated in the KG (P, M).

  • Task 2.2.4: Validate the produced KG against its design requirements. Validation can be performed by the exploitation of the KGs i.e., use KGs in practice/applications e.g., within specific analytics tasks or query formulation based on the CQs (S & P, M).

Process 2.3. Store, publish, query, and visualize the KG. This process is further analyzed in the following tasks:

  • Task 2.3.1: Set up the KG infrastructure in order to host the KG and build the relevant services. This task includes the choice of the software/platform for the KG storage (e.g., Neo4j, OpenLink Virtuoso, RDFox), and the installation and configuration of the software according to the requirements of the KG and its usage (P, M).

  • Task 2.3.2: Store the KG in the developed infrastructure (P, M).

  • Task 2.3.3: Establish query/search interfaces for the stored KG, in order to provide (distributed) KG search services to multiple users, whether they are familiar with query languages (e.g., SPARQL, Cypher) or not (P, M).

  • Task 2.3.4: Establish visualization interfaces for the stored KG. It would be useful to support users with visualization at both levels of knowledge, i.e., the model and data level. These interfaces are useful for the evaluation and deployment tasks. Visualization tasks are aligned with the project goals, resources, and defined requirements. They can also be performed in subsequent iterations (not critical) (P, O).

  • Task 2.3.5: Publish KG in order to make it available to communities of interest and practice which exceed the boundaries of the KG development team but are relevant to the domain of interest. Ideally, these communities may share the same interests and requirements which have been identified initially by the involved team. Publishing the KG should be aligned with the project goals, restrictions, and defined requirements. It can also be performed in subsequent iterations (not critical) (P, O).

3.3 KG evaluation and exploitation phase

The KG evaluation and exploitation phase completes the lifecycle of KG development, including the evaluation of the KG, as well as its deployment and maintenance. Both the evaluation and deployment of the KG may provide valuable feedback for the different processes of the KG development phase and lead to the continuous refinement of the KG in terms of its schema, instance data, and infrastructure (including the various interfaces/tools provided by the infrastructure). Advanced evaluation tasks as described in [46], related to accuracy, coverage, coherency, and succinctness quality aspects, can be conducted to ensure the quality of the created KG. In this phase, the tasks may be performed either individually or conversationally, according to their nature. For example, the measurement of the KG performance may be assigned to a specific member of the team, while the interpretation and improvement of this measurement may be assigned to the whole team or to a subgroup. The recording of individual or conversationally identified issues, comments, and propositions is considered important since it enables the tracking of decisions and changes over different KG versions. The KG evaluation and exploitation phase (Phase 3) includes:

Process 3.1. Evaluation of the quality of the KG, in terms of (i) correctness, (ii) completeness, (iii) bias/fairness (e.g., in sensitive attributes like gender, race). This process is further analyzed in the following tasks:

  • Task 3.1.1: Browse the KG in order to review the most recent version of the KG (S & P, M).

  • Task 3.1.2: Initiate arguments and criticism collaboratively in order to highlight mistakes and propose solutions for the improvement of the KG. This task involves the members of the team that have a clear perception of the KG content and usage, and thus they can identify any deviations from the domain of interest and the requirements of the KG. This task is set as optional since it can be performed in subsequent iterations of the process (not critical) (S, O).

  • Task 3.1.3: Use different metrics in order to measure the quality of KG. The range of coverage, level of detail of the representation and inference possibilities of the KG model, as well as the amount of included data and data sources (by the KG), are aspects that can be measured. Those measurements will be a reference for the comparison of different (even future) versions of the same KG, or different KGs which cover the same domain of interest. This task is set as optional since it can be performed in subsequent iterations of the process (not critical) (P, O).

  • Task 3.1.4: Use the established CQs for the querying/testing of the KG in order to get answers. The use of the CQs over the KG is a best practice for KG testing and discovery of points for improvement. For instance, the whole process of query formulation and answers’ retrieval could emerge issues of query complexity, content incompleteness, etc. This task is recommended; however, it is set as optional since it depends on task 1.2.4. (not critical) (P, O).

  • Task 3.1.5: Compare different versions of the KGs and document the points of similarity and difference. This task may create a useful development and maintenance history for the KG. This task is set as optional since it can be performed in subsequent iterations of the process (not critical) (S, O).

  • Task 3.1.6: Manage recorded issues, exploiting tools for issues’ documentation and sharing between the members of the development team. This task is set as optional since it can be performed in subsequent iterations of the process (not critical) (S, O).

  • Task 3.1.7: Propose new versions (both for the semantic model and the instance data) by incorporating suggested changes. This task is set as optional since it can be performed in subsequent iterations of the process (not critical) (S, O).

  • Task 3.1.8: Define the sensitive attributes (e.g., gender, disability, religion, or ethnicity/race), the sensitive values (e.g., female, blind, Christian, Asian), and the field of potential bias. This is an essential task for mitigating semantic bias of the KG. This task could result to data sources that do not include any sensitive attributes (S, M).

  • Task 3.1.9: Assess bias based on the defined sensitive attributes, values, and general bias field. This task may be performed by analyzing or evaluating the contents of the KG, keeping in mind the pre-defined criteria of bias detection. This task could be omitted only in cases when the data sources do not include any sensitive attributes, or in special scenarios where the bias or the sensitive attributes are going to be studied in the scope of the project (P, O).

Shared space tasks of process 3.1 can be performed by internal (organizational) stakeholders and involved members of the KG engineering team, as well as by external stakeholders, depending on the project goals, constraints, and defined requirements.

Process 3.2. Cleaning of the KG to: (i) improve its correctness, (ii) mitigate bias. This process is further analyzed in the following tasks:

  • Task 3.2.1: Identify wrong assertions of the KG (P, M).

  • Task 3.2.2: Correct wrong assertions of the KG (P, M).

  • Task 3.2.3: Mitigate bias captured in the semantic model. This task could be omitted only in cases when the data sources do not include any sensitive attributes, or in special scenarios where the bias or the sensitive attributes are going to be studied in the scope of the project (P, O).

  • Task 3.2.4: Mitigate bias captured at the instance data. This task could be omitted only in cases when the data sources do not include any sensitive attributes, or in special scenarios where the bias or the sensitive attributes are going to be studied in the scope of the project (P, O).

Process 3.3. Enriching the KG in order to improve the completeness of the KG by adding new statements or improving existing statements. This process is further analyzed in the following tasks:

  • Task 3.3.1: Identify new relevant knowledge sources for the KG. The new sources must meet the requirements that the core sources had met as well (S, O).

  • Task 3.3.2: Apply methods for link discovery between the KG-related sources (P, O).

  • Task 3.3.3: Integrate/merge/align the produced KG with other newly discovered KG sources (P, O).

  • Task 3.3.4: Detect and eliminate duplicates in the enriched KG (P, O).

  • Task 3.3.5: Correct invalid property statements (e.g., domain/range violations) and/or contradicting or uncertain attribute value resolution (in other words, having multiple values for a unique property) (P, O).

All enrichment tasks (Process 3.3) are set as optional since they are aligned with the project goals, resources, and defined requirements. They can also be performed in subsequent iterations (not critical).

Process 3.4. Deploy KG in order to provide services to the involved stakeholders and to the public. This process is further analyzed in the following tasks:

  • Task 3.4.1: Use KG for the development of specific applications by exploiting the structured and processed data that constitute the KG. Such applications include (i) prediction of facts, trends etc., (ii) recommendation of actions, things, etc., (iii) improved query answering/searching, according to the needs of the domain experts or the community that uses the KG after its public provision, (iv) data visualization (S & P, O).

  • Task 3.4.2: Use KG in order to align/merge other relevant KGs for the domain of interest. This task is relevant to KG’s alignment/integration/merging of task 3.3.3, though it entails the provision (design and development) of tools and interfaces for users who are not specialized in the performance of those tasks programmatically. The KG in this case is the base for the development of the tool/service (S & P, O).

All deployment tasks (Process 3.4) are set as optional since they are aligned with the project goals, resources, and defined requirements. They can also be performed in subsequent iterations (not critical).

Process 3.5. Specify maintenance procedures of the KG. This process is further analyzed in the following tasks:

  • Task 3.5.1: Specify the maintenance procedure of the KG, in order to continuously update/refine both the schema and the data from the different sources (S, O).

  • Task 3.5.2: Specify the monitoring procedure of the KG in order to ensure maintenance of high-quality data of the KG (S, O).

All maintenance tasks are highly recommended, and they are set as optional since they are aligned with the project goals, resources, and defined requirements. They can also be performed in subsequent iterations (not critical).

4 Evaluating CHEKG methodology

Motivated by use cases related to drones’ mission for the documentation (with photos) of specific regions and points of interest, we have developed a KG-based approach for transforming trajectories of UAV drones into semantic trajectories (Semantic Trajectory as KG—STaKG) [28]. The semantic trajectories (ST) can be effectively managed, visualized, and analyzed as knowledge graphs using a KG-driven custom-designed toolset. The CHEKG methodology has been applied for the development of STaKGs and the underlying model. The phases of the methodology were adapted and extended by special engineering tasks in order to support the STaKG development.

4.1 Applying the KG specification phase

Based on CHEKG methodology, the involved members of the team were identified and the members’ roles were first specified (Task 1.1.1 and Task 1.1.2). The team included six members: two experts in the field of cartography and geoinformatics, one expert in the field of geoinformatics and software engineering, two ontology engineers and one ontology and software engineer. All members were involved in more than one working group. Particularly, cartography and geoinformatics experts (three members) aimed at the understanding of the domain of interest, provided the main part of the data which constitute the STaKG, established the requirements, and evaluated all the stages and results of design and implementation of the KG. The ontology/knowledge engineers (three members) focused on the design, implementation, and evaluation of the knowledge model, as well as the instantiation of the model with data and eventually the formation of the STaKG. Finally, the knowledge workers (two members) focused on the management of the KG and the design of the tools and services regarding its exploitation and maintenance. The members worked both collaboratively (e.g., shared cloud workspace, Git repositories, WebProtégé, and e-mails) and individually with local documents and tools (Protégé 5.5 and Neo4j).

Afterward, the specification of the aim and scope of the KG followed (Task 1.2.1). In the context of the same CHEKG process (Process 1.2) the data sources (namely, the drones’ log files, information systems that the experts use for the documentation of drone flights, sites, metadata of image files) were identified in collaboration with the domain experts (Task 1.2.2). Also, a set of requirements that the semantic model under development, as well as the KG that integrates it, should satisfy, have been defined (Task 1.2.3). The definition of the requirements has been conducted in collaboration with experts of the domain of interest, and they constitute “points of assessment” for the development and performance of the model and the KG. Finally, the ontology/knowledge engineers and domain experts formulated a set of CQs to be answered against the KG (Task 1.2.4). The specifications and CQs defined in processes 1.1 and 1.2 were documented by the working group for future reference (Task 1.2.5). The process involved interviews with experts in UAV-based documentation and cultural heritage site documentation to gather essential information and knowledge for developing the ontological model and the toolset. This collaborative effort led to the creation of multiple competency questions that engaged all stakeholders. A list of the CQs that were used to evaluate the developed toolset is provided below:

  • Which trajectories of a specific mission include records of a specific object?

  • Which recording positions include records of a specific object?

  • What kind of records are produced during a specific mission?

  • Which missions result in photograph records?

  • What are the recording positions of a specific flight?

  • What kind of records are produced at a specific recording position?

  • What are the recording segments of a trajectory?

  • What are the weather conditions at a specific point in time for a specific flight?

  • Which flights intersect?

  • What is the number of drones involved in a specific mission and the number of flights initiated for that mission?

  • What recording events occurred at a distance of less than 100 m from a specific recording event?

  • Which recording events took place near a specific POI?

The thorough presentation of the defined CQs is included in Kotis et al. [28]

4.2 Applying the KG development phase

At the second phase of CHEKG methodology, during the development of the semantic model, multiple discussions with the domain experts were conducted, to clarify the aspects of the domain of interest and establish a common vocabulary related to the conceptualization that must be developed (Task 2.1.1). Additionally, the data categories which should be correlated and enriched, and eventually supply the KG based on the case study have been determined. The domain experts provided example datasets of the data categories in order to analyze and clean them for further use in the following process. The information extraction process commenced with a physical meeting with the domain experts and the presentation of a) the problem as they view it, b) the data sources that included log files, shapefiles, and image records. Collaborative work was conducted, including discussion to determine the relevant properties and attributes of the data to be extracted for the KG development. Subsequently, the extraction of data from log and image files was programmatically carried out through specialized scripts. The information stored in shapefiles was converted to RDF format, and during the enrichment phase they were retrieved using SPARQL Geospatial queries (Task 2.1.2). In the same context, existing models were identified and studied in order to be reused in the semantic model of the KG (Task 2.1.4). The selection of the models was based on our previous experience in related projects as well as searching ontology repositories (LOV [47] and ODP [48]) for related terms such as trajectory, drone, weather, digitization, recording, record.

Considering the specifications, data analysis, and semantic models research, ontology/knowledge engineers worked on the development of a formal semantic model for the domain of interest, which will constitute the backbone of the KG (Task 2.1.7). The semantic model (Onto4drone [49]) was developed following the HCOME collaborative engineering methodology, supported by Protégé 5.5, and WebProtégé tools. In addition, shared cloud workspaces have been used for further collaborative engineering tasks. It is directly based on the datAcron ontology [45] and indirectly on the DUL, SKOS, SOSA/SSN, SF, GML, and GeoSPARQL ontologies. Additionally, related documentation was added to the developed semantic model (Task 2.1.9), while the result was discussed with the experts i.e., geographers (Task 2.1.10).

As part of the evaluation of the engineered model, as well as for the creation of instance data for the KG the ontology was populated with individuals (Task 2.2.1 and Task 2.2.3). The individuals were part of the data that would constitute the content of the KG, as they have been identified (Task 1.2.2) and gathered (Task 2.1.2) in previous tasks of the work. Additionally, a set of SHACL rules [50] were formulated to evaluate the individuals (Task 2.2.2). Regarding SHACL validation and constraint rule formulation, the Protégé plugin SHACL4Protege [51] was used. Furthermore, for the evaluation of the model and the instance data in this initial stage, the CQs transformed into SPARQL queries via Protégé plugin Snap SPARQL [52] (Task 2.2.4). In the same context, the positions of the UAV were summarized, and only one position was retained per second out of thirty positions per second that were originally tracked (Task 2.2.4). This step effectively reduced the number of positions while maintaining a representative sample of the trajectory of the flight.

Regarding data storage, publishing, retrieval, and visualization, several actions were taken (Phase 2, Process 2.3). The first step was to import the developed ontology into Neo4j using the Neosemantics add-on [53] (Task 2.3.1). Subsequently, the flights’ data, from CSV files, were stored in the graph using Cypher queries and built-in functions (Task 2.3.2). The entities and properties defined in the ontology were used as labels and properties. Additionally, new data from the analysis and enrichment process were also stored in the graph, following the entities, properties, and relations of the ontology.

To facilitate analysis and visualization, the data from the KG were retrieved through manually or automatically generated Cypher queries (Task 2.3.3 and Task 2.3.4). The representation and visualization are achieved through several methods. Spatiotemporal data can be displayed in a tabular form or as points on a map. The records can be represented by the file name, the resource URL, or by displaying the actual record. Furthermore, all data can be visualized as a connected directed graph, allowing for a deeper understanding of the relationships between entities and properties. The source code of developed application that exploits the KG can be found in Github (https://github.com/sotirisAng/stakg) and a live demo can be found at http://stakg-ct-linkdata.aegean.gr.

4.3 Applying the KG evaluation and exploitation phase

The third phase of CHEKG methodology started with the evaluation of the quality of the developed KG (Phase 3, Process 3.1). At this point, the first version of the KG was inspected and discussed in collaboration with the domain experts (Task 3.1.1 and Task 3.1.2). The chosen data sources and respective data have been employed, forming a KG of more than 7K nodes and 10K relations, which derive from the four UAV flight logfiles and enrichment data. The KG was explored and evaluated, in terms of its correctness and completeness, through the set of CQs that experts had provided at the first phase of the development (Task 3.1.1 and Task 3.1.4). The retrievals indicated a few mistakes or omissions regarding the included data, which constituted points for further improvement/refinement. Also, based on the study of the domain of interest of our case (documenting CH POIs recorded from UAVs flights/missions) and the related data, it was concluded that sensitive attributes that could introduce obvious bias were not present (Task 3.1.8).

Moreover, the cleaning process was conducted, in order to ensure that the KG is accurate, consistent, and informative (Phase 3, Process 3.2). Firstly, all temporal data in the log files for the UAV flights and records were converted to UTC format to maintain consistency across the data (Task 3.2.1 and Task 3.2.2). To eliminate duplicates, targeted queries were performed on the graph to identify entities such as records or positions that share identical attributes and subsequently removed them (Task 3.2.1 and Task 3.2.2). Records that did not match a position that was part of a trajectory were excluded, as well as records that lacked location or temporal data (Task 3.2.2). Moreover, positions from the trajectory that had temporal features beyond the timeframe of the flight were removed to ensure that only relevant data were included in the KG (Task 3.2.2).

The KG enrichment process (Phase 3, Process 3.3) involved tasks that aim to improve the completeness of the knowledge graph. One of these tasks was the retrieval of weather data for the specific area and time range of each drone flight and correlating them to the trajectories (Task 3.3.1 and Task 3.3.3). For this task the Historical Weather API [54] was utilized to fetch weather data by sending requests to the API and creating WeatherCondition nodes based on the responses. The created nodes were then connected to trajectory positions in the KG. Another task included in the KG enrichment phase (Task 3.3.1 and Task 3.3.3) was to extract record metadata, which includes geolocation, timestamp, and file name, from the records that are produced during the drone flights. This metadata was then used to define recording events that produce the records. In the enrichment process, external APIs were also utilized to obtain information about points of interest (POIs) documented in OpenStreetMaps [55] or University of the Aegean geographical LOD datasets [56], which might have been recorded during drone flights. This was achieved by the development of methods that form Overpass [57] and SPARQL queries for each record stored in the KG, based on information retrieved from the records. Requests were then sent to the external APIs to execute these queries. This approach enabled the identification of documented POIs located near the drone’s location when the record was produced. This information was then used to merge POI nodes in the knowledge graph and relate them to record nodes. Having developed the first version of the KG, the work focused on the deployment (Task 3.4.1) and specifically on the development and use of a toolset which includes tools for raw trajectory data cleaning, summarization and RDFization, enrichment, semantic trajectory management, and ST browsing and visualization.

The toolset enables the management and retrieval of STaKGs, enrichment of STaKGs, and analysis to recognize semantic behaviors. The raw archive data, which included drone flights’ log files, metadata of recordings, and shapefiles of geographical regions, were semi-automatically annotated by programmatically developed methods, to entities, attributes, and relations based on the Onto4drone ontology. These annotated data were then utilized alongside external open data such as weather data and POIs for the creation and enrichment of STs. The annotation and enrichment processes were based on the ST model, and enriched STs were stored in the Neo4j graph database as a KG. The ST management tool used STaKGs to create trajectory segmentations and perform analytics and tasks, such as merge, split, and combination of trajectories. The web tool for visualization and ST browsing fetched analytics results and STaKG data stored in the KG through predefined and customizable queries to efficiently present them to users who are not specialized in performing those tasks.

Finally, a set of KG maintenance procedures was specified (Task 3.5.1). It includes performing regular enrichment tasks, which involve adding new data to the KG to ensure that it reflects the latest information available. Another maintenance procedure is performing regular cleaning tasks that involve identifying and removing inconsistencies and errors. These tasks follow the enrichment and cleaning processes that are described earlier and are performed to ensure that the KG remains clean and accurate. In addition to them, maintenance procedures for KGs involve updating the KG to align with changes made to the knowledge model/ontology used to structure it (Task 3.5.2). This was achieved through query-based updates and checks of entities, attributes, and relations in the KG, enabling the discovery and elimination of inconsistencies and ensuring that the KG accurately reflects the domain of interest.

5 Discussion and limitations

CHEKG is based on the phases of the ext-HCOME as it supports the decisive involvement of domain experts and knowledge workers (along with knowledge engineers), and it requires the use of tools that facilitate the collaborative management and interaction with conceptualizations in a direct and iterative way, as well as its modularization and bias-assessment/mitigation tasks. When adapting the ext-HCOME for KG engineering, the specification, and conceptualization phases, processes and tasks exhibited an organic and straightforward correspondence. Conversely, certain more specialized tasks involving the actual implementation and utilization of the KG, presented challenges during adaptation and organization into processes and phases, mostly to ensure a comprehensive approach that addresses the diverse aspects of the KG engineering process and KG exploitation. Such tasks included the KG infrastructure development, the creation of interfaces and visualizations for KGs, contextual application descriptions, iterative enrichment, evaluation, and maintenance of KGs.

CHEKG is based on two main advances:

(a) Human-centered and collaborative engineering: CHEKG recognizes the importance of human expertise and domain knowledge in KG development. It emphasizes the active/decisive participation of domain experts and knowledge workers, enabling them to contribute their specialized knowledge and context throughout the process. It provides a systematic and structured approach for involving stakeholders with different levels of expertise and knowledge. Furthermore, it provides personal and shared spaces that facilitate individual and collaborative tasks within the KG development process. These spaces enable domain experts, knowledge workers, and knowledge engineers to work closely together, share their expertise, and iterate on the KG design and content.

(b) Domain-specific specialized tasks and roles: CHEKG recognizes the need for involving domain experts, knowledge engineers, and other stakeholders with specific domain expertise throughout the KG development lifecycle. It defines specialized tasks and roles to ensure that the domain-specific knowledge and requirements are adequately incorporated into the KG. This involvement helps in achieving a more accurate and comprehensive representation of the domain. Moreover, it focuses on the development of fair domain-specific KGs, as it considers the aspects of bias and fairness in KG development.

Existing SotA guidelines and/or methodologies present similarities and correspondences regarding the processes/tasks that must be followed for KG development. The identification of this kind of correspondence for each guideline or methodology has been studied and taken into account for the development of the phases, processes, and tasks of the CHEKG. A mapping of CHEKG phases and existing methodologies is presented in Table 1.

CHEKG is derived from OEMs and integrates specific tasks of ext-HCOME. Phases and processes in CHEKG can also be aligned with those presented in other OEMs. Table 2 provides a mapping of these phases to corresponding ones in collaborative and agile OEMs.

Table 1 Mapping of CHEKG phases to existing methodologies

As emerges from Table 1, CHEKG is drawing parallels with other described methodologies. The detailed and extensive description of the overall KG development process by Fensel et al. [29] and Tamašauskaitė [37] aligns with CHEKG’s systematic structuring. However, these approaches are not necessarily collaborative, and the involved actors are not explicitly described. [39] outlines members of the data ecosystem and emphasizes the role of a knowledge scientist who aims to answer business questions through data exploitation. The XD methodology applied in [23] highlights a collaborative and iterative approach for building KGs, utilizing competency questions. These aspects are also included in CHEKG, which emphasizes on team specification and aim definition along the development and maintenance processes. CHEKG’s alignment with these methodologies showcases its agility and applicability across diverse KG engineering scenarios.

Concerning the fairness and bias mitigation process of the KG, detailed information is missing from the KG methodologies compared to CHEKG. [39] and [36] do not address this issue, [28] states the importance of unbiased data for trustworthiness, and [23] notes that the requirement collection was deliberately template-free and unstructured to prevent bias in the outcome. [23] is the only related work that provides comprehensive coverage of modularity, which adheres to the root-thematic-foundations architectural pattern to construct the ontology underlying the KG. CHEKG highlights the aspect of bias and fairness in tasks 3.1.8, 3.1.9, 3.2.3 and 3.2.4 and the importance of modularity in the creation of explicit knowledge process, particularly in tasks 2.1.1. and 2.1.7.

Table 2 Mapping of CHEKG phases to OEMs

CHEKG methodology was firstly introduced in our preliminary work presented in [25] as STaKG methodology (a predecessor of CHEKG focusing on the engineering of semantic trajectories as knowledge graphs). In the recent work [28], the focus is on the implementation of a domain-specific application, where STaKG methodology was used for the methodological part of engineering ST as KGs. CHEKG aims to provide a generic methodological approach for KG engineering that can be applied across various domains, without relating the engineering of KGs with the engineering of STs only or being tied exclusively to a specific field. CHEKG methodology has been eventually shaped in its latest form by taking into consideration our experience on the entire KG engineering process. It is worth clarifying that a STaKG (a ST represented as a KG) is the output of STaKG methodology, whereas the output of CHEKG methodology can be any type of a KG, including also a STaKG.

The developed KG and the functionalities for its exploitation were able to effectively address the CQs and generate visualizations, as well as to enrich the data by linking them to external sources, as described in Sect. 4.3. The domain experts, who also acted as the end-users of the developed applications, verified the accuracy and practicality of the KG’s content, the correctness of CQs answering, and the efficiency of the visualizations. Furthermore, the KG was published with four showcase datasets, but it requires further expansion with new datasets to examine and validate possible performance and scalability issues. In addition, extra documentation could be added to the semantic model to make it more understandable and reusable.

In the context of the deployment of the methodology, it was noticed that there were a few optional tasks which were i) skipped and ii) applied in a slightly different order. Firstly, the tasks related to bias were omitted for the STaKG development. The domain of interest of our case study (trajectories of UAVs and CH documentation) does not involve sensitive attributes that could introduce obvious bias. However, in future work we will involve experts and conduct proper bias analysis and evaluation. Secondly, some optional tasks were not completed in the proposed order, but they were eventually executed after the development of the exploitation tools. For instance, the enrichment, cleaning, and visualization tasks were performed after the deployment phase due to the availability of the developed tools.

The overall execution time of the project is dependent on the availability of members of the involved team and the project’s time frame. It required approx. 10 man-months for the first iteration and an additional 4 man-months for the second iteration (less than 50 percent of the initial effort). Team members, expressed positive feedback, highlighting the systematic way of work, guided by defined processes and goals. Notably, during the evaluation of CHEKG, it became evident that various team members could negotiate to get the role of the coordinator. While the methodology does not explicitly mention this role, it proves essential, since the team member undertaking this responsibility organizes the frequency, duration, and objectives of meetings, as well as orchestrates the communication pace for different processes and tasks within the team.

CHEKG currently does not provide a comprehensive proposal for tools or techniques specifically tailored for merging KGs. This aspect could be a valuable area for future research and development within the CHEKG framework, exploring effective strategies for aligning and merging KGs.

Although the engineering tasks rely on collaborative tools like Git repositories, cloud-based collaborative workspaces, and engineering tools such as WebProtege, CHEKG currently lacks dedicated tools to actively engage the involved stakeholders in the engineering process. A prospective goal could involve the creation of a specialized toolset to support the methodology, designed to streamline and enhance the efficiency of the overall engineering process.

Considering the assessment of CHEKG, as described in Sect. 4, it is based on a real-world use case scenario. However, to further refine and augment the precision of feedback and to ensure a thorough assessment that extends beyond the immediate practical application, our future objectives include a more formalized approach of the evaluation process. This entails the adoption of a proposed evaluation framework, following the paradigm and mirroring the parameters outlined in related work [24], which could serve as a benchmark for measuring the validity, efficiency, and efficacy of the methodology.

Finally, as was noticed during the use of the methodology, CHEKG provides a collaborative way of working, which could be challenging in some cases, for instance, if the development of the KG involves a single knowledge engineer, or in the case of developing personal KGs. However, even in these cases, the methodology has a flexible structure regarding the tasks that must be followed, allowing the simplification of the required processes.

6 Conclusion and future directions

The role and specific activities followed by the involved team lack detailed description and have not been emphasized in methodological phases and steps. The methodology presented in this paper, namely CHEKG, attempts to fill the aforementioned gaps, following OEM principles, and the main phases of ext-HCOME OE methodology. CHEKG contributes to all phases of the KG engineering lifecycle; it incorporates all the different KG development tasks and organizes them in unambiguous phases, clarifying the roles of the involved members of the development team, exploiting their specialized knowledge in conceptualization, data/knowledge acquisition, KG deployment, and KG evaluation.

So far, CHEKG has been exploited for the development of KGs for the representation of semantic trajectories (STaKG) of drones. The results highlight the feasibility of the methodology, since the team was efficiently organized, and the tasks were fluently conducted. The semantic model, the KG, and the tools developed were positively evaluated by the domain experts who followed the whole process.

Future work includes the usage of the methodology in different domains. Furthermore, the exploitation in the context of different use cases may include working teams with more or less members of varying backgrounds, giving more insights about the efficiency and potential limitations of the methodology in different structures of teams.