Chat with local Llama 3 Model via Ollama in KNIME

#LLMs are still going strong and the development is fast, especially for #opensource and #local LLMs. @mlauber71 explores how to leverage #Llama3 by Meta via the #Ollama wrapper and prompt the model as a #lowcode chatbot using #KNIME. Enjoy the data story!

PS: :date:#HELPLINE . Want to discuss your article? Need help structuring your story? Make a date with the editors of Low Code for Data Science via Calendly → Calendly - Blog Writer

3 Likes

#LLMs can be customized to provide smarter responses using user-curated knowledge bases and adopting #RAG. @mlauber71 uses #Llama3 locally via the #Ollama wrapper in #KNIME to also build a #vectorstore with his own PDFs, CSVs or log files. Enjoy the data story!

PS: If you’re on #KNIME 5.1 or higher, you can leverage the power of local #LLMs, including Llama 3 via Ollama, easily and swiftly with the #KNIME #AI extension. :wink:

PS 2: :date:#HELPLINE . Want to discuss your article? Need help structuring your story? Make a date with the editors of Low Code for Data Science via Calendly → Calendly - Blog Writer

2 Likes