Solutions to “Just KNIME It!” Challenge 18 - Season 3

:boom: A new Just KNIME It! challenge just came out on explainable AI (xAI). :mag: Let’s explore the LIME technique with this data puzzle to make research on breast cancer prediction models more transparent. :stethoscope:

Here is the challenge. Let’s use this thread to post our solutions to it, which should be uploaded to your public KNIME Hub spaces with tag JKISeason3-18 .

:sos: Need help with tags? To add tag JKISeason3-18 to your workflow, go to the description panel in KNIME Analytics Platform, click the pencil to edit it, and you will see the option for adding tags right there. :blush: Let us know if you have any problems!

4 Likes

Here’s my solution. Rather than trying to completely reinvent the wheel, I used the LIME workflow from the Hub. It required a lot of revision, but seems to work fine. Probably could be improved with some parameter optimization.



5 Likes

My solution to the challange:

I have created a neural network for the prediction

  • Input layer: 8 nodes
  • Hidden layer: 64 nodes
  • Hidden layer: 32 nodes
  • Hidden layer: 16 nodes
  • Output layer: 2 nodes (softmax for the probabilities)

I created the LIME chart for all the test observations (averaged the individuals), and this is my final chart:

I think this challange rightly tagged as a hard challange, but was very exciting for me!

5 Likes

Hello all,
Here is my solution

You can find actual feature values impact using tile view

5 Likes

Hi all,
Here is my solution. I used a simple random forest without any optimizations. I calculated the average LIME value for each class, as well as the LIME values for each sample, to determine which features might contribute to the predictions. I believe that optimizing the prediction conditions could alter these results. Thank you.


5 Likes

Hi all,
Here is my solution. I created a generic LIME analysis workflow for the binary classification model. This workflow will work for different datasets as long as the appropriate target column is selected in the Column Settings component.


In this workflow, you can focus on False Positives and False Negatives to analyze which features contributed to the incorrect predictions. This will be useful for improving the model and understanding the quality of the data.


6 Likes

Find herewith my submission - JKISeason3-18 – KNIME Community Hub

3 Likes

Hi all,
Here is my solution.

3 Likes

Hi, this is my solution for this challenge.

I modified two workflows from community hub.

3 Likes

Getting one on the board as well… with most machine learning topics still a bit out of my depth, but amazed how one can pull things off by following examples.

In order not to just copy any other solution I set up parameter optimization for Gradient Boosted Tree and then wrangled the data so that I could use some of the visuals from the LIME Example on the hub (Bubble and Violine plot), plus some of the bar charts that were used here.

2 Likes

:sun_with_face: Another Tuesday, another solution to our beloved Just KNIME It! challenges! :sun_with_face:

:exploding_head: Once again you folks went above and beyond with the modeling and visualization for this challenge. Bravo!! :clap:

:arrow_down: Here’s our solution highlighting the power of LIME.

:wind_face: :wind_face: See you tomorrow for a fast challenge on Formula 1 analytics!

3 Likes