As anyone who processes large text collections would be aware, the TF node can be painfully slow to execute, to the extent that it can become a major bottleneck in my workflow, and I sometimes revert to using DF instead. (As I write this, the TF node has been chewing through a collection of 11,000 news articles for several minutes, and is not even half done.) I assume that the slowness results from every document being processed in succession rather than in parallel. I make this assumption because the processing can be sped up considerably by containing the node in a parallel chunk loop.
Given that many other text processing nodes have inbuilt parallel processing options, is there any reason why the TF node does not? And parallel processing aside, might there be any room for other efficiency improvements in this node?