From Dynamic to Evolvable Knowledge Graphs in Manufacturing: Systematic Literature Review on Learning Approaches

Tracking #: 3745-4959

Authors: 
Anna Teern
Nada Elgendy
Markus Kelanti
Henna Tammia
Tero Päivärinta

Responsible editor: 
Guest Editors KG Construction 2024

Submission type: 
Survey Article
Abstract: 
This systematic literature review investigates how evolvable KGs enhance manufacturing under Industry 4.0 and 5.0, identifying key technologies and gaps. Evolvable KGs adapt to changing knowledge by leveraging machine learning algorithms and human expertise to enhance decision-making, operational efficiency, and predictive maintenance capabilities beyond the capabilities of dynamic KGs. Despite the advancements, challenges persist in quality assurance, process planning, and the integration of human expertise. The findings advocate for addressing these issues to foster wider adoption and optimization of KG technologies in manufacturing. This review maps existing literature based on the KG construction process stages. The main results are the state-of-the-art tasks in creating evolvable KGs in manufacturing and categories of learning approaches in evolvable KGs. The results contribute by updating the KG construction process and widening the understanding of evolvable KGs that utilize learning. By deepening the understanding of how KGs can evolve, this review sets the base for future research to develop more dynamic and intelligent systems tailored to the emerging demands of Industry 5.0.
Full PDF Version: 
Tags: 
Reviewed

Decision/Status: 
Major Revision

Solicited Reviews:
Click to Expand/Collapse
Review #1
Anonymous submitted on 01/Jan/2025
Suggestion:
Major Revision
Review Comment:

The paper provides a systematic literature review about learning approaches for updating knowledge graphs (KG) in the manufacturing domain. While the topic of the paper is in general interesting and relevant for the community, it fails to present the content clearly and comprehensibly. Thus, I suggest to improve the entire article by (1) rethinking the overall alignment and (2) reworking the presentation of the results.

The strong points of the papers are:
S1: The paper addresses an interesting and relevant topic.
S2: The paper is based on a sound research method.

The following opportunities for improvement could be identified:
O1: The terms and distinction between “dynamic” and “evolve” KG is not clear.
O2: The knowledge graph construction process (KGCP) should be reconsidered, and the results of the paper should be aligned with the single steps for a clear presentation.
O3: The Introduction should be rewritten to clearly outline the content of the paper.
O4: The classification of learning method (Tables 7, 8, 9) should be reconsidered.
O5: Data quality assurance in KGs is represented insufficiently.

O1-O5 are detailed in the following paragraphs.

O1: The definitions and distinction between “dynamic” and “evolve” KG are obviously introduced by the authors themselves and should not be assumed to be known. Thus, I either suggest to (i) define and differentiate both terms more clearly right at the beginning of the paper – by also citing others work than the ones provided by the authors – or (ii) remove the distinction from the paper. I personally think that the paper would not lose anything if the distinction between the two terms is removed. The terms are especially confusing in title and abstract and do not really explain what can be expected from the paper. A title like “A Systematic Literature Review on Knowledge Graph Construction in Manufacturing” would seem more appropriate to me.

O2: Knowledge Graph Construction Process.
Here, purely addressing the „construction“ of KGs seems a bit limited since the authors also discuss the evolvability beyond an initially constructed KG, that is, the KG lifecycle. Second, the single steps should be numbered and a distinction between the ontology and instances should be made. This structure should then be used as structure for Section 3. Currently, it is not clear, why the structure of Section 3 does not align with the KGCP. Instead of purely building on the author’s own work, I suggest taking a broader look into KG literature and rethinking the KKCP:
- Zhong, L., Wu, J., Li, Q., Peng, H., & Wu, X. (2023). A comprehensive survey on automatic knowledge graph construction. ACM Computing Surveys, 56(4), 1-62.
- Paulheim, H. (2017). Knowledge graph refinement: A survey of approaches and evaluation methods. Semantic Web, 8(3), 489-508.

O3: Introduction.
The current Introduction does not outline what can be expected from the paper. I suggest moving the outline from page 3 before subsections 1.1 and 1.2. It does not quite get clear why the authors suddenly introduce the KGCP without motivating its use for the survey. In addition, the concepts of dynamic and evolvable KG (see O1) are used without definition. I suggest heavily reworking the Introduction by (1) clearly framing the context, (2) providing an overview on existing research, and (3) stating the research gap -> which leads to the authors contribution.

O4: Classification of learning methods.
Table 7, 8, and 9 basically provide one of the core contributions of this work, namely, a classification of the identified learning methods. Unfortunately, the organization in these classes is neither explained nor intuitive. For example, in Table 7, it is not clear why reinforcement learning appears first as subclass of Representation, and second as separate class below representation. In Table 8, it is not clear whether the approaches address only changes in the ontology or also within the instances. Also, the categorization in Table 9 is not clear. It seems that the authors confused the “what is updated” (i.e., data/knowledge) with “who caused the update” (i.e., humans vs. machines). While “human input”, “cooperation”, and “machine-enriched updates” make sense, mixing in “data/knowledge updates” needs more explanation. To me, it seems that this is a separate aspect and should not be confused with the cause or trigger of the change. Further, captions are missing for the columns: the relationship between the first and second column in Tables 7 and 8 is not clear. It would also be helpful to highlight which papers consider one or more learning approaches, i.e., if a very specific one is used very often or if many papers use a wide variety of learning approaches. This aspect is not addressed at all.

O5: Data quality in knowledge graphs.
While the authors presume that “the quality of knowledge within KGs is an overlooked aspect” (cf. 4.3 Challenges) they do not sufficiently consider related work about this aspect. One reason might be the unclear assignment of quality assurance within the KGCP, which leads to a spread discussion of the topic throughout the paper: some parts are discussed in 3.3.1. Cleaning as a type of data preprocessing, some aspects are discussed in 3.5, and some other aspects in 3.7.1 Changes through inferencing algorithms – Node changes. A global look at the topic of quality assurance in KGs would be useful here. I suggest the authors to consider the following papers (amongst others) and rethink their integration of KG quality within the KGCP:
- Rabbani, K., Lissandrini, M., & Hose, K. (2023). Extraction of validating shapes from very large knowledge graphs. Proceedings of the VLDB Endowment, 16(5), 1023-1032.
- Issa, S., Adekunle, O., Hamdi, F., Cherfi, S. S. S., Dumontier, M., & Zaveri, A. (2021). Knowledge graph completeness: A systematic literature review. IEEE Access, 9, 31322-31339.

Further comments:
- Introduce all abbreviations before the first use (e.g., KG in the abstract, DT, OPC UA).
- Explain the concepts of Industry 4.0 and 5.0 shortly in the Introduction with one sentence, not assuming this should be of general knowledge by all readers.
- Introduction: I suggest using the original cite for the linked data paper mentioned in 1.1.
- Introduction: triples (noun-verb-noun): here, subject-predicate-object is most probably more common.
- Research questions should be refined such that all three of them refer to the singular or plural consistently. E.g., for RQ2 it is unclear whether task automation is investigated within one solution or in general for all observed solutions.
- Page 4: “The context is (IKG) …” -> remove parentheses.
- Standardize automated vs. automatized.
- EC6: clarify how poor quality is determined.
- Table 4: information on how many papers were excluded per EC should be added. In addition, it should be mentioned whether the exclusion criteria were applied in the respective order or whether more than one could be assigned to one result.
- TRL levels: it would have been interesting how many results where excluded due to EC8 having a TRL<3. For basic research papers, lower TRL levels would have been expected, but in my perception, these works would still yield very interesting solutions for the paper to be investigated. Especially wrt. to new trends of KG learning that are not yet implemented in productive KGs. If possible, a revised version should include those works as well.
- Section 3 should be renamed to “Results”.
- The Google Knowledge Graph should be cited or footnoted.
- Fig. 4, 5, and 6 should be ordered by decreasing number of publications. In addition, for each Figure it should be mentioned whether a publication has a single assignment or multiple possible ones.
- 3.1. Ontologies: here, I miss a discussion that in real-world manufacturing settings, automatically built ontologies typically do not meet the quality standards and expectations of the domain experts out of the box.
- Section 3.4.: “For the ontology, Protégé is certainty the most used tool …” -> since the section is about storage of knowledge, it is not clear whether Protégé is used for storing or building the ontology. The latter is mor probably assumed.
- Page 14: “data or knowledge updates” -> clarify whether data or knowledge is meant.
- Page 14: the citation for evolvable KGs should be properly cited.
- Page 14: “… the human gives their …” -> unclear whether singular or plural.
- Cite SKOS and explain abbreviation.
- Page 17: “… to allow some conclusions.” -> which conclusions?
- Discussion: a bullet point list with the most important findings would be useful to make the contribution better readable.
- Conclusion: “In conclusion, evolvability and related learning approaches …” -> so far, the concept of evolvability was defined through learning approaches, this sentence is rather confusing.
- The references should be reviewed since many journal names are missing and first names are written inconsistency (sometimes full first name, sometimes abbreviated). Examples: [33, 67, 73]

Review #2
Anonymous submitted on 24/Mar/2025
Suggestion:
Major Revision
Review Comment:

This systematic review examines how manufacturing knowledge graphs evolve by surveying 38 studies to map their lifecycle—from ontology selection and automated knowledge extraction to integration and inferencing and identifies four modes of KG learning (human input, human–machine cooperation, data‑driven updates, and machine‑enriched updates). It finds that most industrial KG solutions remain mid‑TRL (4–5), are mainly focused on adding new knowledge via link prediction and largely neglect removal of outdated information and rigorous quality evaluation.

The manuscript asserts the importance of KGs for Industry 4.0/5.0, yet never clearly distinguishes this concept from existing “dynamic” or “temporal” KG paradigms. On page 2, paragraph 3, the authors write that “Evolving knowledge or KGs are usually described in terms of temporal variations of knowledge” but fail to cite or critically contrast more recent work on temporal KGs. This ambiguity undermines the reader’s ability to understand what novel capability “evolvability” adds. To improve clarity, the authors should insert a concise definition, contrasting static, dynamic, and evolvable KGs with precise citations.

The paper describes its search strategy in prose and although Tables 1 through 3 show some information related to search terms and exclusion criteria, this information might not be sufficient to replicate this work. Furthermore, the exclusion of studies below TRL 3 is stated without enough justification. The authors should explain why proof‑of‑concept work is irrelevant for mapping the field’s trajectory. Moreover, although Cohen’s kappa values are reported for screening agreement, the process for resolving reviewer conflicts is described only superficially. A paragraph detailing how discrepancies were adjudicated would strengthen methodological transparency.

The literature that is reviewed, catalogues KG construction tasks exhaustively (Fig.4), but it lacks a critical synthesis of comparative performance or maturity. For example, Tables 7–9 list link‑prediction algorithms without reporting accuracy, computational cost, or industrial applicability. Readers are left with a descriptive taxonomy rather than insight into which approaches have proven effective in real manufacturing deployments. An improvement would be the inclusion of a discussion that compares reported results based on metrics used in literature, that links back to TRL.

The authors correctly identify gaps (e.g.few methods remove obsolete triples) in the discussion section, but they stop short of diagnosing root causes or proposing concrete research directions. Phrases such as “This is problematic…” appear without follow‑through recommendations. A more scientifically coherent discussion could evaluate why removal of outdated knowledge remains rare (e.g.complexity of graph versioning) and propose specific methodological remedies drawn from related domains.

Finally, the authors restate the survey findings without translating them into actionable guidance for practitioners or clear next steps for researchers. It fails to propose a roadmap, by for instance, defining a minimal evaluation framework for KG quality or suggesting benchmarking for dynamic KG evaluation. By explicitly articulating how evolvable KG design principles could evolve (e.g.integrating incremental embedding updates, user‑feedback loops, and provenance tracking), the conclusion could fulfil its promise to shape future work rather than merely summarize it.

Review #3
Anonymous submitted on 26/Mar/2025
Suggestion:
Minor Revision
Review Comment:

This paper presents a literature review about evolvable KGs in the manufacturing domain, and focuses on how KGs evolve when integrating ML techniques and human interaction.
The paper is very well written and clear.

My main concern is that the authors selected papers addressing industry/manufacturing KGs, with a special focus on how the KGs have been constructed (methodology), and they discuss in general how different tasks in KG construction (e.g., knowledge acquisition, cleaning) have been addressed. However, it seems to me that, by selecting only the manufacturing domain, they may be missing important work about similar/different methodologies for KG construction applied to different domains, but they often talk about KG construction tasks in general, with no actual discussion about how some tasks have peculiar characteristics when applied to the manufacturing domain (see for example section 3.3.1). So, I am not sure how useful part of this literature review is in practice. E.g., if I need to understand how I can select ontologies to reuse, I would rather go and read a literature review that in general addresses this topic, not one that limits its discussion to papers developing an ontology for manufacturing. Instead, I find more useful sections like 3.6 or 3.7.2, about how KGs are used in manufacturing processes. Also, the remarks included in the Discussion section are useful.

More specific comments below:
- The extraction of the articles is a bit outdated (snowballing is dated Sept 2023, one year before the paper submission).
- At page 6 you talk about "the original review": I wanted to understand if this work (dated March 2022) has been already published somewhere (and, if so, if it's all included in [69]), or it's just a work that has been suspended for some time and then resumed in late 2023.
- It's not clear to me from Table 2 how the authors select only those articles that talk about "evolvable KGs"
- Where can I find the complete list of selected papers? If it's not available anywhere, it should be. Also because it's not always clear which references in the paper are part of the study, and which ones are just cited for other reasons.
- Figure 4: I would explain better in the paper all the tasks listed in the figure. Also, what the "..." stand for in "Ontology construction/..."?
- Section 3.1: "In this study, KGs consist of ontologies..." I would clarify this before in the paper
- Section 3.1: "The knowledge acquisition stage includes selecting an ontology...": from this section, it seems that ontologies are part of knowledge acquisition (and in section 4.1 it is said that ontology construction activities are placed within the knowledge acquisition stage), but then the following section is about "knowledge acquisition" that is described as a different thing: please, clarify this. Also, I would like these sections to be more consistent: "knowledge acquisition" is a task, while "ontologies" are not a task alone, right?
- Section 3.3: Wouldn't it be relevant to talk more about ontology matching?