I had installed the 4.3 version and it was getting worse and worse in efficiency during two days.
I have installed 4.5 to see if it could improve. I have a two flows in the same workflow with just about 50 nodes of grouping, combining columns, filtering, pivoting, nothing special. If I reset the nodes the whole workflow runs in about a minute, not more.
Once finished if I just try to click on a node knime begins to consume 70% CPU and doesn’t open the node for configuration until 15-20 seconds have passed
Intel i7 series 8, 16GB ram, 6GB reserved for Knime memory.
Thanks for your quick response,
Yes is a Win11
Two additional info:
I’m parsing a Json file (30 MB) Json read–>Json Path–>Ungroup
It seems that now something is going wrong because after ungrouping 30.000 items correctly the rest is totally disorganized this could be causing huge files… I guess
On the other hand, sometimes Knime works well but after opening several times the result tables to audit the results things begin to go super slow, just to continue opening those tables.
I will read your suggested posts, even if I know some of them already,
@Gonzo I have no experience with Windows 11 as of now.
These things come to my mind when you mention a large ungroup (or any other complex operation for that matter):
the Cache node. Sometimes I have experienced that before complex operations after a long data manipulation stream KNIME ‘likes’ a Cache node in order to bring all the changes back to one place
then you could see if forcing the node to write something to disk can help, also if you try different underlying encodings in the knime.ini (-Dknime.compress.io=[SNAPPY|GZIP|NONE]) of the workflows, maybe try GZIP instead of SNAPPY (this might come at a cost in speed and will be the default for every workflow), Also see this entry.
try the columnar table backend. KNIME has not yet moved it to a standard but you can activate it for individual workflows
and then you could think about splitting the ungroup in chunks (If you data does allow that)
A comibination of these along with some more RAM and garbage collection (1 | 2 ) at the start maybe might bring you across the finish line
I’m using a “Json path” node, as my data is already a list I have to pass it through that node before ungrouping.
The “Json path” node comes by default with a “$…*” parameter, I just kept that untouched. The result is that the node worked… almost well, so no error but not well.
Just writing “$." instead of "$…” has solved the problem.
In addition the output tables that in origin had 1,3Mlines were multiplied x6 number of lines, now they stay clean.
I’l update if I’ve solved also the memory problem, in any case, it should be less critical
Hi @Gonzo , it may be a good idea to understand what the JSON path statements mean.
The "$..*" means take everything. I think Knime adds this by default so that if you don’t do anything inside the node, the node simply acts as a pass-through. Similarly with nodes such as Python script, Java script, etc, by default Knime will output what the nodes are receiving as input, so that they act as pass-through in case you don’t do anything.
That being said, if you do not need all the data, then do not use the default $..* path. Use only the paths that you need.
Make sure you understand what you are doing, especially if you are going to re-use the workflow with another input. While this is working with your current data, it might not work with another data if you are going by trial and error.