Review Comment:
Decision: Major Revision
## High-level Review
I would like to thank the authors for this submission. The paper summarizes existing XAI techniques and conducted a survey to find possible applications of XAI techniques in the domain of knowledge graph engineering. It is an easy to digest paper for introducing XAI to the knowledge engineering community. The following are the feedbacks that I have for this submission
## General feedback
- There is a previous work by the author which covers 1/3 of the content in
this paper
[https://doi.org/10.3233/FAIA230091](https://doi.org/10.3233/FAIA230091)
word for word. If this journal is an extension upon this previous paper,
there is a need for clear indication in the introduction section: i)
clarifying what this work contributes on top of the previous work (the
delta), ii) that this work is a clear extension of the previous work and not
a standalone work.
- In my opinion, this paper could be evaluated better as a **survey paper**
rather than a full research paper. A majority of this work focuses on
studying a significant amount of related literatures (specifically other
systematic review papers on XAI's) to analyze them, summarize/categorize
them, derive new findings from those existing works and conduct an interview
to verify the findings. Otherwise, if evaluated as a full research paper,
this paper, as is, lacks the originality of a full research paper since it
is not clearly stated and the results of the interview study are
insignificant to push new boundaries, it just made the requirements for XAI
methods in knowledge graph engineering clearer. Thus, I cannot recommend
accepting this paper in its current form as a full research paper.
- In the methodology section for literature review, the use cases are defined
without any motivation nor citations on why these particular sets of use
cases are used for analysis of the collected literatures. How are these use
case derived from the literature? Are there any literatures to support these
use cases in terms of AI usage in knowledge graph constructions?
There are also a few problems with the provided use cases. Although the
following two problems are handled in the interview study methodology, they
are not mentioned in the literature review methodology. Firstly, I do not
see any mentions of regulations compatibility as part of the use case
analysis even though it was a significant focus in the problem context
paragraph of the introduction section. Is it possible to also consider
regulations compatibility as part of the analysis of the literatures?
Secondly, there is no mention of data provenance considerations in the use
cases, which is a significant part of the problem context described in the
introduction section (this is also related to the previous point about
regulations compatibility).
- For the pool of interview participants, it would be interesting if there
were more participants from the industry (currently, there are only 3
industry participants) leading to the results of the interview study leaning
more to the academia.
- Explanation Design (Figure 6) and Section 4.4.2 only gave details on the
requirement analysis for **users**, **use cases**, and **representations of
explanations**. A major component on **regulations** was just briefly
mentioned in the paper and **not discussed in depth** leaving it as just a
background context for the introduction section of the paper. Similarly,
"Evaluation" step is also left out from the in-depth discussion in Section
4.4.2.
Regarding the XAI design blueprint step on "Evaluation", are there
recommendations for the selection process on the type of metrics/dimensions
when evaluating the XAI models? If this is mentioned somewhere in the
"Findings" sections, it would be nice to reiterate the recommendations when
describing the proposed XAI design blueprint.
- It would also be nice to indicate where/which part of the findings answers
the 4 research questions mentioned in the introduction section.
- A few citations lack either a DOI or a URL to the paper. It would be nice to
have URLs to follow the citations.
## Detailed Review
### Introduction
**Page 2:**
- KG lifecycle --> what do you mean with KG lifecycle? The construction phase?
The usage of KG? The storage of KG? (Rereading it, it is fully introduced in
Background 2.4, but it would be nice to have a short sentence explaining
what it is in introduction)
- Most regulators take a risk-based approach to the use of AI ... are
compliant with the law (line 10-13) --> is there a citation to support these
two statements?
- Up-to-date comparative surveys... (line 20-21) --> this statement is very
disconnected from the rest of the paragraph on human-centric approach.
Remove it if it's not needed.
- ... we would like to advance the field of **explainable knowledge
engineering** (line 23-24) --> How? A sentence or two to show "how" would
strongly support this statement. If the "how" is development of
human-in-the-loop approaches for transparency and accountability, the
accompanying sentence needs to be rewritten/restructured for more clarity.
### Background
Page 4:
- Reviews and surveys ... from **end-users** have also become increasingly
common (line 11-12) --> Looks very out of place since the paragraph is
mostly focused on XAI without involvement of end-users. Would also need
citations if you decide to keep this statement.
Page 5:
- KGs are interacting with AI capabilities in complex ways (line 30) -->
How/What are the complex ways AI interacts with KG in the figure? From the
figure it looks pretty simple since the **input** for stage C, where I
assume most of the AI methods/models are, comes from stage D and as
**output**, it enriches the generated KG. Similarly, stage D also has a very
clear input/output direction.
- While KGs constructed using these approaches ... similar transparency
challenges as the algorithms it complements (line 45-47) --> Doesn't it mean
that crowdsourcing approach, in general, is a bad idea since it results in
biased, bad quality data while also suffering from transparency issues? This
sentence doesn't read well.
### Methodology
Page 9:
- Table 3's tasks is not aligned with the tasks mentioned in Stage B of Figure
1). Is this intentional? If yes, it would be nice to also have another
column with relevant tasks that are aligned with the ones provided in the
figure for knowledge graph construction stage of the KG lifecycle.
Page 9-10:
- 3.2.1 Interview questions section (2 paragraphs)
The first paragraph leads the reader through the interview process
step-by-step until the end where risk concerns are addressed. It reads well
and has an _order_ to it. However, the second paragraphs came in totally
disconnected talking about the "examples" selection, which I believe is for
the topics **Use Cases**, **XAI Example Discussion**, and **Requirements**
of the interview.
I think it would read better to make the second paragraph a separate
subsection titled "Examples and use cases selection from literature process"
and link the sentences "Inspired by \[55\], we designed ... concerns,
challenges, and requirements" (Page 9 line 45-47) to that section.
### Findings
#### SOTA study/review
Page 12:
- a human-in-the-loop system that complies ... (line 41) --> compiles (do you
mean compile?)
- For instance, NERO uses... (line 51) --> ... NERO \[++citation\] (citation
missing)
Page 13
- SIRE employs... (line 20) --> SIRE \[++citation\] (citation)
- Beyond NERO, LogiRE ...(line 24) --> LogiRE \[++citation\] (citation)
- Diverging from text-based explanations, ProtoRE ...(line 24) --> ... ProtoRE
\[++citation\] (citation)
- RULESYNTH, proposed by Singh et al.... (line 35) --> citations reference
link?
- Last sentence on _Entity Resolution_: Additionally, ... attempt to add them
to make non-matching pairs more similar. It took me a while to read this
sentence and understand it. If it is about entity pairs which are
different/non-matching, but **contextually** similar due to the input
attributes, this sentence needs to be rewritten to provide the clarity.
Page 14:
- which require feeding more data and extending training time. (line 31) -->
which require feeding more data **thus** extending training time (it reads
better this way?)
* The relationship between the complexity of functions and ... educate them
--> What kind of relationship? Complex functions + more freedom of
operations leads to lesser time required to educate the users?
- approxSemanticCrossE proposed explanation... target the link --> targetting?
#### Use cases and capabilities
Page 15:
- Among the use cases, three areas, ... (line 42-43) -> Which three areas?
- such as explainers designed for any knowledge... and some mode-specific
methods... -> such explainers designed for **both** any knowledge ... and ?
|