Hi KNIME Support team.
I have a memory question regarding Pyspark Script.
After creating a Spark Session by connecting with Livy, I am importing about 5 million data into parquet and running ML analysis code with Pyspark Script.
-
After completing the Pyspark Script, where is the data stored in memory? Is it stored in Spark Driver memory?
-
If it is stored in a specific space, is it possible to reduce the memory by clearing the cache in that specific space?
Your answers will be appreciated.