2021-10-12 16:15:57,818 : ERROR : KNIME-Worker-26-Create Databricks Environment 0:2214 : : Node : Create Databricks Environment : 0:2214 : Execute failed: Problem with reading the data, class org.knime.bigdata.databricks.rest.libraries.ClusterLibraryStatus, ContentType: application/json. javax.ws.rs.client.ResponseProcessingException: Problem with reading the data, class org.knime.bigdata.databricks.rest.libraries.ClusterLibraryStatus, ContentType: application/json. at org.apache.cxf.jaxrs.impl.ResponseImpl.reportMessageHandlerProblem(ResponseImpl.java:469) at org.apache.cxf.jaxrs.impl.ResponseImpl.doReadEntity(ResponseImpl.java:411) at org.apache.cxf.jaxrs.client.AbstractClient.readBody(AbstractClient.java:554) at org.apache.cxf.jaxrs.client.ClientProxyImpl.handleResponse(ClientProxyImpl.java:979) at org.apache.cxf.jaxrs.client.ClientProxyImpl.doChainedInvocation(ClientProxyImpl.java:867) at org.apache.cxf.jaxrs.client.ClientProxyImpl.invoke(ClientProxyImpl.java:308) at com.sun.proxy.$Proxy60.getClusterStatus(Unknown Source) at org.knime.bigdata.spark.core.databricks.context.DatabricksClient.findJobJarOnCluster(DatabricksClient.java:287) at org.knime.bigdata.spark.core.databricks.context.DatabricksClient.installJobJarOnCluster(DatabricksClient.java:224) at org.knime.bigdata.spark.core.databricks.context.DatabricksSparkContext.uploadJobJar(DatabricksSparkContext.java:325) at org.knime.bigdata.spark.core.databricks.context.DatabricksSparkContext.open(DatabricksSparkContext.java:222) at org.knime.bigdata.spark.core.context.SparkContext.ensureOpened(SparkContext.java:145) at org.knime.bigdata.spark.core.databricks.node.create.DatabricksSparkContextCreatorNodeModel2.executeInternal(DatabricksSparkContextCreatorNodeModel2.java:143) at org.knime.bigdata.spark.core.node.SparkNodeModel.execute(SparkNodeModel.java:240) at org.knime.core.node.NodeModel.executeModel(NodeModel.java:556) at org.knime.core.node.Node.invokeFullyNodeModelExecute(Node.java:1259) at org.knime.core.node.Node.execute(Node.java:1039) at org.knime.core.node.workflow.NativeNodeContainer.performExecuteNode(NativeNodeContainer.java:559) at org.knime.core.node.exec.LocalNodeExecutionJob.mainExecute(LocalNodeExecutionJob.java:95) at org.knime.core.node.workflow.NodeExecutionJob.internalRun(NodeExecutionJob.java:201) at org.knime.core.node.workflow.NodeExecutionJob.run(NodeExecutionJob.java:117) at org.knime.core.util.ThreadUtils$RunnableWithContextImpl.runWithContext(ThreadUtils.java:365) at org.knime.core.util.ThreadUtils$RunnableWithContext.run(ThreadUtils.java:219) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at org.knime.core.util.ThreadPool$MyFuture.run(ThreadPool.java:123) at org.knime.core.util.ThreadPool$Worker.run(ThreadPool.java:246) Caused by: com.fasterxml.jackson.databind.exc.InvalidFormatException: Cannot deserialize value of type `org.knime.bigdata.databricks.rest.libraries.LibraryInstallStatus` from String "SKIPPED": not one of the values accepted for Enum class: [INSTALLING, RESOLVING, INSTALLED, UNINSTALL_ON_RESTART, FAILED, PENDING] at [Source: (sun.net.www.protocol.http.HttpURLConnection$HttpInputStream); line: 1, column: 634] (through reference chain: org.knime.bigdata.databricks.rest.libraries.ClusterLibraryStatus["library_statuses"]->java.util.ArrayList[1]->org.knime.bigdata.databricks.rest.libraries.LibraryStatus["status"]) at com.fasterxml.jackson.databind.exc.InvalidFormatException.from(InvalidFormatException.java:67) at com.fasterxml.jackson.databind.DeserializationContext.weirdStringException(DeserializationContext.java:1851) at com.fasterxml.jackson.databind.DeserializationContext.handleWeirdStringValue(DeserializationContext.java:1079) at com.fasterxml.jackson.databind.deser.std.EnumDeserializer._deserializeAltString(EnumDeserializer.java:327) at com.fasterxml.jackson.databind.deser.std.EnumDeserializer._fromString(EnumDeserializer.java:214) at com.fasterxml.jackson.databind.deser.std.EnumDeserializer.deserialize(EnumDeserializer.java:188) at com.fasterxml.jackson.databind.deser.impl.FieldProperty.deserializeAndSet(FieldProperty.java:138) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:324) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:187) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer._deserializeFromArray(CollectionDeserializer.java:347) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:244) at com.fasterxml.jackson.databind.deser.std.CollectionDeserializer.deserialize(CollectionDeserializer.java:28) at com.fasterxml.jackson.databind.deser.impl.FieldProperty.deserializeAndSet(FieldProperty.java:138) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:324) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:187) at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:322) at com.fasterxml.jackson.databind.ObjectReader._bind(ObjectReader.java:2007) at com.fasterxml.jackson.databind.ObjectReader.readValue(ObjectReader.java:1174) at com.fasterxml.jackson.jaxrs.base.ProviderBase.readFrom(ProviderBase.java:816) at org.apache.cxf.jaxrs.utils.JAXRSUtils.readFromMessageBodyReader(JAXRSUtils.java:1409) at org.apache.cxf.jaxrs.impl.ResponseImpl.doReadEntity(ResponseImpl.java:402) ... 25 more 2021-10-12 16:15:57,822 : WARN : KNIME-Worker-26-Create Databricks Environment 0:2214 : : AbstractFvvalIfSwitchNodeModel2 : IF Switch (Flow Variable Value) : 0:2160:2173:1688 : No valid variable selected 2021-10-12 16:15:57,822 : WARN : KNIME-Worker-26-Create Databricks Environment 0:2214 : : Node : IF Switch (Flow Variable Value) : 0:2160:2173:1688 : No valid variable selected 2021-10-12 16:15:57,822 : WARN : KNIME-Worker-26-Create Databricks Environment 0:2214 : : Node : JSON Path : 0:2160:2173:1686:1672 : No suitable (JSON) column found: name=default,columns=[]