Hi Nathan, welcome to the forum. Thanks for bringing up this issue.
@victor_palacios, this could be related indeed. This is currently on my list for investigation. I’ll let you know once I have more information.
Hi Nathan, welcome to the forum. Thanks for bringing up this issue.
@victor_palacios, this could be related indeed. This is currently on my list for investigation. I’ll let you know once I have more information.
Hi Victor (and @BenjaminMoser thanks for your feedback ,here is some more detail from the log file.
I cannot share the workflow file, but you will burst out laughing when you see the workflow picture (it could not be simpler). This is not the same as I was using with my original work, but to make your bug hunt easier I replicated the errors using the Kaggle Titanic Dataset (train.csv).
If there was a way for me to share non-publicly (perhaps a quick zoom call? that would probably be preferable, and if so please contact me directly).
What I notice when I look at the log files
Network paths partially suppressed in logs, but it is not C:
Sadly our network directories have spaces and & in them.
This decision predates me.
Log file for Tableau Writer
2021-12-22 10:40:23,654 : ERROR : KNIME-Worker-4-Tableau Writer 0:2 : : Node : Tableau Writer : 0:2 : Execute failed: The database "hyper.file:\\***********************\KNIME%20Dev\Data\Titanic\tableau_writer.hyper" could not be created. Hyper was unable to resolve the database file path "\\*****************\KNIME%20Dev\Data\Titanic": Directory does not exist
Context: 0x5fdfad59
com.tableau.hyperapi.HyperException: The database "hyper.file:\\************************\KNIME%20Dev\Data\Titanic\tableau_writer.hyper" could not be created. Hyper was unable to resolve the database file path "\\***********************************\KNIME%20Dev\Data\Titanic": Directory does not exist
Context: 0x5fdfad59
at com.tableau.hyperapi.Connection.<init>(Connection.java:177)
at com.tableau.hyperapi.Connection.<init>(Connection.java:78)
at org.knime.ext.tableau.hyperapi.TableauHyperAPIProcess$HyperAPIConnection.<init>(TableauHyperAPIProcess.java:210)
at org.knime.ext.tableau.hyperapi.TableauHyperAPIProcess.createConnection(TableauHyperAPIProcess.java:125)
at org.knime.ext.tableau.hyperapi.TableauHyperWriter.<init>(TableauHyperWriter.java:90)
at org.knime.ext.tableau.hyperapi.write.TableauHyperWriterNodeModel2.execute(TableauHyperWriterNodeModel2.java:133)
at org.knime.core.node.NodeModel.execute(NodeModel.java:758)
at org.knime.core.node.NodeModel.executeModel(NodeModel.java:549)
at org.knime.core.node.Node.invokeFullyNodeModelExecute(Node.java:1267)
at org.knime.core.node.Node.execute(Node.java:1041)
at org.knime.core.node.workflow.NativeNodeContainer.performExecuteNode(NativeNodeContainer.java:559)
at org.knime.core.node.exec.LocalNodeExecutionJob.mainExecute(LocalNodeExecutionJob.java:95)
at org.knime.core.node.workflow.NodeExecutionJob.internalRun(NodeExecutionJob.java:201)
at org.knime.core.node.workflow.NodeExecutionJob.run(NodeExecutionJob.java:117)
at org.knime.core.util.ThreadUtils$RunnableWithContextImpl.runWithContext(ThreadUtils.java:365)
at org.knime.core.util.ThreadUtils$RunnableWithContext.run(ThreadUtils.java:219)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at org.knime.core.util.ThreadPool$MyFuture.run(ThreadPool.java:123)
at org.knime.core.util.ThreadPool$Worker.run(ThreadPool.java:246)
Log File for Parquet Writer when the configuration is set to:
Mountpoint and then Local
2021-12-22 10:40:21,077 : ERROR : KNIME-Worker-3-Parquet Writer 0:3 : : Node : Parquet Writer : 0:3 : Execute failed: /\\**********************************\KNIME Dev\Data\Titanic\parquet_writer.parquet
java.nio.file.NoSuchFileException: /\\**********************************\KNIME Dev\Data\Titanic\parquet_writer.parquet
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystem.toLocalPathWithAccessibilityCheck(LocalWorkflowAwareFileSystem.java:311)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystem.getEntity(LocalWorkflowAwareFileSystem.java:205)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.getEntity(LocalWorkflowAwareFileSystemProvider.java:121)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.checkSupport(LocalWorkflowAwareFileSystemProvider.java:112)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.newOutputStreamInternal(LocalWorkflowAwareFileSystemProvider.java:126)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.newOutputStreamInternal(LocalWorkflowAwareFileSystemProvider.java:1)
at org.knime.filehandling.core.connections.base.BaseFileSystemProvider.newOutputStream(BaseFileSystemProvider.java:500)
at java.base/java.nio.file.Files.newOutputStream(Unknown Source)
at org.knime.bigdata.hadoop.filesystem.NioFileSystem.create(NioFileSystem.java:151)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:910)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:891)
at org.apache.parquet.hadoop.util.HadoopOutputFile.create(HadoopOutputFile.java:74)
at org.apache.parquet.hadoop.ParquetFileWriter.<init>(ParquetFileWriter.java:329)
at org.apache.parquet.hadoop.ParquetWriter.<init>(ParquetWriter.java:292)
at org.apache.parquet.hadoop.ParquetWriter$Builder.build(ParquetWriter.java:658)
at org.knime.bigdata.fileformats.parquet.ParquetFileFormatWriter.<init>(ParquetFileFormatWriter.java:112)
at org.knime.bigdata.fileformats.parquet.ParquetFormatFactory.getWriter(ParquetFormatFactory.java:172)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.createWriter(FileFormatWriter2NodeModel.java:289)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.writeToFile(FileFormatWriter2NodeModel.java:257)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.write(FileFormatWriter2NodeModel.java:214)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:185)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:1)
at org.knime.core.node.NodeModel.executeModel(NodeModel.java:549)
at org.knime.core.node.Node.invokeFullyNodeModelExecute(Node.java:1267)
at org.knime.core.node.Node.execute(Node.java:1041)
at org.knime.core.node.workflow.NativeNodeContainer.performExecuteNode(NativeNodeContainer.java:559)
at org.knime.core.node.exec.LocalNodeExecutionJob.mainExecute(LocalNodeExecutionJob.java:95)
at org.knime.core.node.workflow.NodeExecutionJob.internalRun(NodeExecutionJob.java:201)
at org.knime.core.node.workflow.NodeExecutionJob.run(NodeExecutionJob.java:117)
at org.knime.core.util.ThreadUtils$RunnableWithContextImpl.runWithContext(ThreadUtils.java:365)
at org.knime.core.util.ThreadUtils$RunnableWithContext.run(ThreadUtils.java:219)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at org.knime.core.util.ThreadPool$MyFuture.run(ThreadPool.java:123)
at org.knime.core.util.ThreadPool$Worker.run(ThreadPool.java:246)
Log file extract when Parquet is set to Local
2021-12-22 10:47:08,087 : ERROR : KNIME-Worker-11-Parquet Writer 0:3 : : Node : Parquet Writer : 0:3 : Execute failed: C:\**************************\KNIME Dev\Data\Titanic\parquet_writer.parquet
java.nio.file.NoSuchFileException: C:\*************************\KNIME Dev\Data\Titanic\parquet_writer.parquet
at java.base/sun.nio.fs.WindowsException.translateToIOException(Unknown Source)
at java.base/sun.nio.fs.WindowsException.rethrowAsIOException(Unknown Source)
at java.base/sun.nio.fs.WindowsException.rethrowAsIOException(Unknown Source)
at java.base/sun.nio.fs.WindowsFileSystemProvider.newByteChannel(Unknown Source)
at org.knime.filehandling.core.fs.local.fs.LocalFileSystemProvider.newByteChannel(LocalFileSystemProvider.java:148)
at java.base/java.nio.file.spi.FileSystemProvider.newOutputStream(Unknown Source)
at java.base/java.nio.file.Files.newOutputStream(Unknown Source)
at org.knime.bigdata.hadoop.filesystem.NioFileSystem.create(NioFileSystem.java:151)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:910)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:891)
at org.apache.parquet.hadoop.util.HadoopOutputFile.create(HadoopOutputFile.java:74)
at org.apache.parquet.hadoop.ParquetFileWriter.<init>(ParquetFileWriter.java:329)
at org.apache.parquet.hadoop.ParquetWriter.<init>(ParquetWriter.java:292)
at org.apache.parquet.hadoop.ParquetWriter$Builder.build(ParquetWriter.java:658)
at org.knime.bigdata.fileformats.parquet.ParquetFileFormatWriter.<init>(ParquetFileFormatWriter.java:112)
at org.knime.bigdata.fileformats.parquet.ParquetFormatFactory.getWriter(ParquetFormatFactory.java:172)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.createWriter(FileFormatWriter2NodeModel.java:289)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.writeToFile(FileFormatWriter2NodeModel.java:257)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.write(FileFormatWriter2NodeModel.java:214)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:185)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:1)
at org.knime.core.node.NodeModel.executeModel(NodeModel.java:549)
at org.knime.core.node.Node.invokeFullyNodeModelExecute(Node.java:1267)
at org.knime.core.node.Node.execute(Node.java:1041)
at org.knime.core.node.workflow.NativeNodeContainer.performExecuteNode(NativeNodeContainer.java:559)
at org.knime.core.node.exec.LocalNodeExecutionJob.mainExecute(LocalNodeExecutionJob.java:95)
at org.knime.core.node.workflow.NodeExecutionJob.internalRun(NodeExecutionJob.java:201)
at org.knime.core.node.workflow.NodeExecutionJob.run(NodeExecutionJob.java:117)
at org.knime.core.util.ThreadUtils$RunnableWithContextImpl.runWithContext(ThreadUtils.java:365)
at org.knime.core.util.ThreadUtils$RunnableWithContext.run(ThreadUtils.java:219)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at org.knime.core.util.ThreadPool$MyFuture.run(ThreadPool.java:123)
at org.knime.core.util.ThreadPool$Worker.run(ThreadPool.java:246)
Image of workflow also attached.
Nathan
@NKlassen out of curiosity. Have you tried using the new Path variables? You could construct the filename in a Java node and convert it to a variable that the Parquet writer might use and see if that makes any difference:
And can you tell us what kind of file system it is you are writing to - what kind of shared directory it is?
Hello @mlauber71
Thanks for the suggestion.
I have not yet tried that, but this worked in 4.4 without a workaround.
I’ll give it a shot in the interim.
I do not have much insight into the architectural backend of the organization I work for.
So I’m not sure what I can provide that would be helpful.
It’s an NTFS shared drive that, to the best of my knowledge, is accessed by mostly Windows users.
It is accessilble to me as either \\Domain Name\Myriad\of\Directories
or
Driver_Letter:\Myriad\of\Directories
@victor_palacios @BenjaminMoser
I did try to replicate this locally at home over Christmas on my personal computer on a local drive using the same version (4.5) of KNIME.
I tried a variety of combinations of nested directories, directory names with spaces and symbols on them on my Windows 10 desktop. I didn’t encounter any issues on my local drive.
So this may be a network folder specific issue.
Nathan
Hi all,
My intuition tells me this is due to how these nodes handle spaces in UNC paths (i.e. on shared drives).
This is on my list to look into. I think I can do this by the end of the week. Chances are good that the next release will contain a fix for this issue, but I can’t make any promises right now.
I’ll let you know once I have further details.
Thanks for all the support. @NKlassen, thanks for offering a call to elaborate. I don’t think it is necessary right now but I’ll get back to you if I have a need.
Best,
Benjamin
My pleasure!
Just curious if there was a change in UNC related functionality between 4.4 and 4.5 that may have caused this?
Please let me know if I can provide any other support!
Take care!
@NKlassen Not to UNC paths per se (as far as I’m aware), but we did make changes to how special characters in paths are handled.
Hi,
I had a look at this and a fix is currently awaiting review (AP-18190 for internal reference).
I did not really have the chance to think much about finding a workaround for the intermediate time. Here’s one thing that worked for my simple test case. However, you may run into authentification issues or other problems.
The idea is to create a symbolic link to the problematic folder s.t. the link name does not contain spaces.
In the windows command prompt (you may need admin privileges), you can create a symbolic directory link in the current working directly with mklink /d link-target-name "\\host\path\files to read"
.
This creates a new link called link-target-name
that points at the shared directory. You can then set the node to read from this path, e.g. C:\Users\Knime\Desktop\files-to-read-link
instead of \\host\path\files to read
.
Best Regards,
Ben
Thanks @BenjaminMoser for the update!
I’ll give this work around a test in the near future and let you know if it works
Thanks for all your help!
Nathan
Hi @NKlassen,
the Parquet nodes do not support UNC paths, but there is a new SMB Connector node to connect Windows network things. Does the new connector node fix the problem with Parquet?
Cheers,
Sascha
Hi all,
a potential workaround that is probably safer and easier is to map the share (i.e. the UNC path) to a local drive. In the node configuration, you should be able to select the target file/directory on the newly mapped drive just like it would be your good old C:\
.
Cheers,
Ben
Thank you!
I also was finally able to find a ‘traditional’ filepath, not UNC path, which worked for the node in the interim as well.
Still looking forward to a future fix so I can implement the same way I do for other writer nodes
Thank you for everyone’s help,
Nathan
@NKlassen I just realised you had already mentioned in an earlier post that you can access the directory via a drive letter. That’s exactly what I had had in mind when proposing another possible workaround yesterday.
In any case, we have a bugfix release coming up next week and chances are good that it will contain a change that improves the handling of UNC paths with special characters such as spaces. This will potentially fix your issue with the Tableau Writer.
Cheers,
Ben
Hi,
a few days ago we released 4.5.1 which should fix the regressions with the Tableau writer. You can download the new version here.
Cheers,
Ben
Hi @BenjaminMoser ,
Did some testing using 4.5.1 - same workflow as before.
The Tableau Writer works, even with a UNC path!
Unfortunately, the Parquet Writer does not work with a UNC path.
Parquet Writer still requires the specific driver letter and associated path.
The actual node has the following directory in it and is set to Local File System.
\\DOMAIN\Area\Silly Directory Name with & and Spaces\Another Silly Directory with Spaces\KNIME Dev\Data\Titanic\test.parquet
The error messages shows the following but at the C:… which is not the appropriate drive letter.
ERROR Parquet Writer 4:9 Execute failed: C:\Area\Silly Directory Name with & and Spaces\Another Silly Directory with Spaces\KNIME Dev\Projects\Bug Reporting\test.parquet
The error messages are different when the node is set to Mount Point and LOCAL
ERROR Parquet Writer 4:6 Execute failed: /\\DOMAIN\Area\Silly Directory Name with & and Spaces\Another Silly Directory with Spaces\KNIME Dev\Data\Titanic\parquet_writer_mtpoint.parquet
Please let me know how/if I can help.
Nathan
Forgot to post the logs, apologies @BenjaminMoser
Local File System Setting for Parquet Node
2022-01-25 11:49:24,074 : ERROR : KNIME-Worker-26-Parquet Writer 4:9 : : Node : Parquet Writer : 4:9 : Execute failed: C:\Area\Silly Directory with & and Spaces\Another Directory with Spaces\KNIME Dev\Projects\Bug Reporting\test.parquet
java.nio.file.NoSuchFileException: C:\Area\Silly Directory with & and Spaces\Another Directory with Spaces\KNIME Dev\Projects\Bug Reporting\test.parquet
at java.base/sun.nio.fs.WindowsException.translateToIOException(Unknown Source)
at java.base/sun.nio.fs.WindowsException.rethrowAsIOException(Unknown Source)
at java.base/sun.nio.fs.WindowsException.rethrowAsIOException(Unknown Source)
at java.base/sun.nio.fs.WindowsFileSystemProvider.newByteChannel(Unknown Source)
at org.knime.filehandling.core.fs.local.fs.LocalFileSystemProvider.newByteChannel(LocalFileSystemProvider.java:148)
at java.base/java.nio.file.spi.FileSystemProvider.newOutputStream(Unknown Source)
at java.base/java.nio.file.Files.newOutputStream(Unknown Source)
at org.knime.bigdata.hadoop.filesystem.NioFileSystem.create(NioFileSystem.java:146)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:910)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:891)
at org.apache.parquet.hadoop.util.HadoopOutputFile.createOrOverwrite(HadoopOutputFile.java:81)
at org.apache.parquet.hadoop.ParquetFileWriter.<init>(ParquetFileWriter.java:327)
at org.apache.parquet.hadoop.ParquetWriter.<init>(ParquetWriter.java:292)
at org.apache.parquet.hadoop.ParquetWriter$Builder.build(ParquetWriter.java:658)
at org.knime.bigdata.fileformats.parquet.ParquetFileFormatWriter.<init>(ParquetFileFormatWriter.java:112)
at org.knime.bigdata.fileformats.parquet.ParquetFormatFactory.getWriter(ParquetFormatFactory.java:172)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.createWriter(FileFormatWriter2NodeModel.java:289)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.writeToFile(FileFormatWriter2NodeModel.java:257)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.write(FileFormatWriter2NodeModel.java:214)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:185)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:1)
at org.knime.core.node.NodeModel.executeModel(NodeModel.java:549)
at org.knime.core.node.Node.invokeFullyNodeModelExecute(Node.java:1267)
at org.knime.core.node.Node.execute(Node.java:1041)
at org.knime.core.node.workflow.NativeNodeContainer.performExecuteNode(NativeNodeContainer.java:559)
at org.knime.core.node.exec.LocalNodeExecutionJob.mainExecute(LocalNodeExecutionJob.java:95)
at org.knime.core.node.workflow.NodeExecutionJob.internalRun(NodeExecutionJob.java:201)
at org.knime.core.node.workflow.NodeExecutionJob.run(NodeExecutionJob.java:117)
at org.knime.core.util.ThreadUtils$RunnableWithContextImpl.runWithContext(ThreadUtils.java:367)
at org.knime.core.util.ThreadUtils$RunnableWithContext.run(ThreadUtils.java:221)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at org.knime.core.util.ThreadPool$MyFuture.run(ThreadPool.java:123)
at org.knime.core.util.ThreadPool$Worker.run(ThreadPool.java:246)
Mountpoint / Local Settings for Parquet Node
2022-01-25 11:51:36,378 : ERROR : KNIME-Worker-27-Parquet Writer 4:6 : : Node : Parquet Writer : 4:6 : Execute failed: /\\DOMAIN\Area\Silly Directory with & and Spaces\Another Directory with Spaces\KNIME Dev\Data\Titanic\parquet_writer_mtpoint.parquet
java.nio.file.NoSuchFileException:/\\DOMAIN\Area\Silly Directory with & and Spaces\Another Directory with Spaces\KNIME Dev\Data\Titanic\parquet_writer_mtpoint.parquet
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystem.toLocalPathWithAccessibilityCheck(LocalWorkflowAwareFileSystem.java:311)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystem.getEntity(LocalWorkflowAwareFileSystem.java:205)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.getEntity(LocalWorkflowAwareFileSystemProvider.java:121)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.checkSupport(LocalWorkflowAwareFileSystemProvider.java:112)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.newOutputStreamInternal(LocalWorkflowAwareFileSystemProvider.java:126)
at org.knime.filehandling.core.fs.knime.local.workflowaware.LocalWorkflowAwareFileSystemProvider.newOutputStreamInternal(LocalWorkflowAwareFileSystemProvider.java:1)
at org.knime.filehandling.core.connections.base.BaseFileSystemProvider.newOutputStream(BaseFileSystemProvider.java:500)
at java.base/java.nio.file.Files.newOutputStream(Unknown Source)
at org.knime.bigdata.hadoop.filesystem.NioFileSystem.create(NioFileSystem.java:146)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:910)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:891)
at org.apache.parquet.hadoop.util.HadoopOutputFile.createOrOverwrite(HadoopOutputFile.java:81)
at org.apache.parquet.hadoop.ParquetFileWriter.<init>(ParquetFileWriter.java:327)
at org.apache.parquet.hadoop.ParquetWriter.<init>(ParquetWriter.java:292)
at org.apache.parquet.hadoop.ParquetWriter$Builder.build(ParquetWriter.java:658)
at org.knime.bigdata.fileformats.parquet.ParquetFileFormatWriter.<init>(ParquetFileFormatWriter.java:112)
at org.knime.bigdata.fileformats.parquet.ParquetFormatFactory.getWriter(ParquetFormatFactory.java:172)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.createWriter(FileFormatWriter2NodeModel.java:289)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.writeToFile(FileFormatWriter2NodeModel.java:257)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.write(FileFormatWriter2NodeModel.java:214)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:185)
at org.knime.bigdata.fileformats.node.writer2.FileFormatWriter2NodeModel.execute(FileFormatWriter2NodeModel.java:1)
at org.knime.core.node.NodeModel.executeModel(NodeModel.java:549)
at org.knime.core.node.Node.invokeFullyNodeModelExecute(Node.java:1267)
at org.knime.core.node.Node.execute(Node.java:1041)
at org.knime.core.node.workflow.NativeNodeContainer.performExecuteNode(NativeNodeContainer.java:559)
at org.knime.core.node.exec.LocalNodeExecutionJob.mainExecute(LocalNodeExecutionJob.java:95)
at org.knime.core.node.workflow.NodeExecutionJob.internalRun(NodeExecutionJob.java:201)
at org.knime.core.node.workflow.NodeExecutionJob.run(NodeExecutionJob.java:117)
at org.knime.core.util.ThreadUtils$RunnableWithContextImpl.runWithContext(ThreadUtils.java:367)
at org.knime.core.util.ThreadUtils$RunnableWithContext.run(ThreadUtils.java:221)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at org.knime.core.util.ThreadPool$MyFuture.run(ThreadPool.java:123)
at org.knime.core.util.ThreadPool$Worker.run(ThreadPool.java:246)
Hi @NKlassen,
thanks again for the detailed feedback.
As indicated above, the Parquet nodes do not officially support UNC paths. For the time being, you can look into Sascha’s suggestion to use the SMB Connector node or try one of the workarounds I suggested.
Cheers,
Ben
Apologies Ben.
I misunderstood Sashca’s point, I had thought it was a temporary issue with Parquet.
Thanks again for the clarification and both of your workarounds!
Have a nice day!
This topic was automatically closed 7 days after the last reply. New replies are no longer allowed.