There is no explicit limit but a JVM string cannot be bigger than 2G. It will also at some point run out of memory with too big of a query plan tree or become incredibly slow due to query planning complexity. I've seen queries that are tens of MBs in size.

On Thu, Jul 11, 2019 at 5:01 AM, 李书明 <> wrote:
I have a question about the limit(biggest) of SQL's length that is supported in SparkSQL. I can't find the answer in the documents of Spark.

Maybe Interger.MAX_VALUE or not ?