Multimodal prompting with local LLMs using KNIME and Ollama

What if your text and images could be analyzed by the same LLM entirely offline? :light_bulb: @mlauber71 shows how to use KNIME + Ollama to enable multimodal prompting with local models like mistral-small 3.1 and analyze car insurance claims and images β€” secure, private, and powerful. :rocket: Enjoy the data story!

PS: :date:#HELPLINE . Want to discuss your article? Need help structuring your story? Make a date with the editors of Low Code for Data Science via Calendly β†’ Calendly - Blog Writer

3 Likes

There seems to be a discrepancy between the post and the actual model output for one of the accidents.

1 Like

@rfeigel good catch :slight_smile: will correct this. Not sure an actual insurance CEO will use this model. This was to demonstrate that the use of text and images in a local LLM is possible even on a local machine. Actually implementing them in production requires some effort and in might involve using a major LLM secured against data leaks and hosted in a region where you have legal protection.

Very nice project. I think its a good demo of multimodal capabilities. A production model would have to handle geographic variability of labor rates. To do some crude testing, you could add a locale to the accident description. I live in the US. In my case you could compare New Jersey (high cost) with Vermont (relatively lower cost.) Such a comparison would do some testing of the LLM’s β€œblackboxiness”.