Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarify more the research ideas for text and images #187

Closed
elboyran opened this issue Jan 13, 2023 · 5 comments
Closed

Clarify more the research ideas for text and images #187

elboyran opened this issue Jan 13, 2023 · 5 comments
Assignees

Comments

@elboyran
Copy link
Contributor

elboyran commented Jan 13, 2023

The proposal sates the following generic research question:

Can generic properties of the XAI relevance scores be identified by systematic study?

and

Not much research on the difficult “explaining the explainer” problem is done  focus on intuitive
insights from binary classification on simple datasets.
in Scope:

Applied research on XAI methods 1-3 (RISE, LIME
and KernelSHAP) properties on binary classification
of simple image, text and time-series datasets

Make it more concrete, create hypothesis and design initial experiments.

@elboyran elboyran self-assigned this Jan 13, 2023
@elboyran
Copy link
Contributor Author

elboyran commented Jan 16, 2023

Practicalities

  • Log in experiments and results in dianna-exploration, Issues also there?

  • Focus on text and then images

  • Use only our simple test datasets:

Stanford sentiment (text) It is possible to browse the dataset and filter it, but it seems positive is encoded blue, while negative is in red.

and

Simple shapes (images). Use the generation script to define test images for the experiments!

  • Start with RISE, then LIME and KernalSHAP

  • Since we decided not to implement 2 data items side by side in the DIANNA dashboard, one should either use 2 browser tabs with 2 running dashboards or Notebooks for the experiments.

@elboyran elboyran changed the title Clarify more the research tasks and pitch to the team Clarify more the research tasks for text (and images) and pitch to the team Jan 17, 2023
@elboyran
Copy link
Contributor Author

elboyran commented Jan 17, 2023

Hypothesis

With fixed model and (default, demo) model parameters, relevance maps change consistently with the change of input independent of the data modality.

Experiments

Text

  • Design lists of 25 positive and negative words using the sentiment scale of the Stanford sentiment treebank.
  • Create 25 test sentences of length 3 containing the above words, one sentence each. Perhaps as simple as

This is terrible
This is great
This is marvelous

etc.
Make them similar to, but different from the training data (see the link above), we'll use them for testing.

- Testing the hypothesis
For each XAI method
For a set of continuously changing method parameters
Evaluate:

  • Qualitatively: use 2 dashboard instances to to compare pairs of sentences and their relevance maps. Evaluate the relevance score strength of the words of interest. See issue Explore text explanations dianna#439.
  • Quantitatively: use a notebook (?) to display the relevancy vs the sentiment scale (per class?).

Images

  • Use the generation script to create ~10 test images per shape per shape parameter. Shape parameters are:
  • for circles: size and gray-scale contrast? (BG and FG)
  • for triangles: size, rotation angle and gray-scale contrast? (BG and FG)

- Testing the hypothesis
For each XAI method
For a set of continuously changing method parameters
For each shape parameter (with others fixed)
Evaluate:

  • Qualitatively: use 2 dashboard instances to compare pairs of "neighboring" (in the variable shape parameter) shapes and their relevance maps. Evaluate the (average/mean/max/median?) relevance score strength. See issue Explore text explanations dianna#439.
  • Quantitatively: use a notebook (?) to display the (average/mean/max/median?) relevancy vs the shape parameter scale (per shape = class?).

@elboyran elboyran changed the title Clarify more the research tasks for text (and images) and pitch to the team Clarify more the research tasks for text and images Jan 17, 2023
@elboyran elboyran changed the title Clarify more the research tasks for text and images Clarify more the research ideas for text and images Jan 31, 2023
@elboyran
Copy link
Contributor Author

@loostrum would you like to "review" my ideas?

@loostrum
Copy link
Member

I think these are some nice, concrete tasks that we could work on. For the quantitative work indeed it may be easiest to do this with notebooks. When running the dashboard by hand, we should make sure to store our results from that somehow. Perhaps just make screenshots of interesting results?

@elboyran
Copy link
Contributor Author

When running the dashboard by hand, we should make sure to store our results from that somehow. Perhaps just make screenshots of interesting results?

Good point, I'll add this to the actual issues description.

@elboyran elboyran transferred this issue from dianna-ai/dianna Jun 13, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants