Hi All -
I’ve been digging into the code and looking into what appears to be a memory leak (https://jira.apache.org/jira/browse/SPARK-27337) and have noticed something kind of peculiar about the way closing a SparkSession is handled. Despite being marked as Closeable, closing/stopping a SparkSession simply stops the SparkContext. This changed happened as a result of one of the PRs addressing https://jira.apache.org/jira/browse/SPARK-15073 in https://github.com/apache/spark/pull/12873/files#diff-d91c284798f1c98bf03a31855e26d71cR596.
I’m trying to understand why this is the intended behavior – anyone have any knowledge of why this is the case?