From 59aa26b4c5bb095ab5e62cd86ec9adc2d3181894 Mon Sep 17 00:00:00 2001 From: Shri Saran Raj N Date: Tue, 10 Dec 2024 20:40:14 +0530 Subject: [PATCH] Handle default value for maxExecutors sparkConf Signed-off-by: Shri Saran Raj N --- .../src/main/scala/org/apache/spark/sql/FlintJobExecutor.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/spark-sql-application/src/main/scala/org/apache/spark/sql/FlintJobExecutor.scala b/spark-sql-application/src/main/scala/org/apache/spark/sql/FlintJobExecutor.scala index d0fb201a6..135cc6dbd 100644 --- a/spark-sql-application/src/main/scala/org/apache/spark/sql/FlintJobExecutor.scala +++ b/spark-sql-application/src/main/scala/org/apache/spark/sql/FlintJobExecutor.scala @@ -548,7 +548,7 @@ trait FlintJobExecutor { } def getSegmentName(sparkSession: SparkSession): String = { - val maxExecutorsCount = sparkSession.conf.get(FlintSparkConf.MAX_EXECUTORS_COUNT.key) + val maxExecutorsCount = sparkSession.conf.get(FlintSparkConf.MAX_EXECUTORS_COUNT.key, "unknown") String.format("%se", maxExecutorsCount) }