WARN Chat Model Prompter 4:133 Traceback (most recent call last): File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_transports\default.py", line 101, in map_httpcore_exceptions yield File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_transports\default.py", line 250, in handle_request resp = self._pool.handle_request(req) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_sync\connection_pool.py", line 256, in handle_request raise exc from None File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_sync\connection_pool.py", line 236, in handle_request response = connection.handle_request( ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_sync\http_proxy.py", line 288, in handle_request connect_response = self._connection.handle_request( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_sync\connection.py", line 101, in handle_request raise exc File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_sync\connection.py", line 78, in handle_request stream = self._connect(request) ^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_sync\connection.py", line 124, in _connect stream = self._network_backend.connect_tcp(**kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_backends\sync.py", line 207, in connect_tcp with map_exceptions(exc_map): File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpcore\_exceptions.py", line 14, in map_exceptions raise to_exc(exc) from exc httpcore.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 996, in _request response = self._client.send( ^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_client.py", line 914, in send response = self._send_handling_auth( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_client.py", line 942, in _send_handling_auth response = self._send_handling_redirects( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_client.py", line 979, in _send_handling_redirects response = self._send_single_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_client.py", line 1014, in _send_single_request response = transport.handle_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_transports\default.py", line 249, in handle_request with map_httpcore_exceptions(): File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\httpx\_transports\default.py", line 118, in map_httpcore_exceptions raise mapped_exc(message) from exc httpx.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Program Files\KNIME\plugins\org.knime.python3.nodes_5.4.1.v202501291500\src\main\python\_node_backend_launcher.py", line 1055, in execute outputs = self._node.execute(exec_context, *inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\plugins\org.knime.python3.nodes_5.4.1.v202501291500\src\main\python\knime\extension\nodes.py", line 1237, in wrapper results = func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\plugins\org.knime.python.llm_5.4.2.v202502111701\src\main\python\src\models\base.py", line 735, in execute answer = _call_model_with_output_format_fallback( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\plugins\org.knime.python.llm_5.4.2.v202502111701\src\main\python\src\models\base.py", line 757, in _call_model_with_output_format_fallback return response_func(model) ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\plugins\org.knime.python.llm_5.4.2.v202502111701\src\main\python\src\models\base.py", line 733, in get_response return chat.invoke(conversation_messages) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\langchain_core\language_models\chat_models.py", line 158, in invoke self.generate_prompt( File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\langchain_core\language_models\chat_models.py", line 560, in generate_prompt return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\langchain_core\language_models\chat_models.py", line 421, in generate raise e File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\langchain_core\language_models\chat_models.py", line 411, in generate self._generate_with_cache( File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\langchain_core\language_models\chat_models.py", line 632, in _generate_with_cache result = self._generate( ^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\langchain_openai\chat_models\base.py", line 522, in _generate response = self.client.create(messages=message_dicts, **params) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_utils\_utils.py", line 279, in wrapper return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\resources\chat\completions.py", line 863, in create return self._post( ^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 1283, in post return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 960, in request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 1020, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 1098, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 1020, in _request return self._retry_request( ^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 1098, in _retry_request return self._request( ^^^^^^^^^^^^^^ File "C:\Program Files\KNIME\bundling\envs\org_knime_python_llm_5.4.2\Lib\site-packages\openai\_base_client.py", line 1030, in _request raise APIConnectionError(request=request) from err openai.APIConnectionError: Connection error. ERROR Chat Model Prompter 4:133 Execute failed: Connection error.