Skip to content

Commit

Permalink
ut
Browse files Browse the repository at this point in the history
  • Loading branch information
turboFei committed Jan 11, 2024
1 parent fdc1155 commit 367622b
Showing 1 changed file with 50 additions and 0 deletions.
Original file line number Diff line number Diff line change
@@ -0,0 +1,50 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/

package org.apache.spark.sql.kyuubi

import org.apache.spark.sql.execution.{CollectLimitExec, TakeOrderedAndProjectExec}
import org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec
import org.apache.spark.sql.internal.SQLConf

import org.apache.kyuubi.engine.spark.WithSparkSQLEngine

class SparkDatasetHelperSuite extends WithSparkSQLEngine {
override def withKyuubiConf: Map[String, String] = Map.empty

test("get limit from spark plan") {
Seq(true, false).foreach { aqe =>
val topKThreshold = 3
spark.sessionState.conf.setConf(SQLConf.ADAPTIVE_EXECUTION_ENABLED, aqe)
spark.sessionState.conf.setConf(SQLConf.TOP_K_SORT_FALLBACK_THRESHOLD, topKThreshold)
spark.sql("CREATE OR REPLACE TEMPORARY VIEW tv AS" +
" SELECT * FROM VALUES(1),(2),(3),(4) AS t(id)")

val topKStatement = s"select * from tv order by id limit ${topKThreshold - 1}"
var sparkPlan = spark.sql(topKStatement).queryExecution.sparkPlan
assert(sparkPlan.isInstanceOf[TakeOrderedAndProjectExec] || sparkPlan.asInstanceOf[
AdaptiveSparkPlanExec].inputPlan.isInstanceOf[TakeOrderedAndProjectExec])
assert(SparkDatasetHelper.planLimit(sparkPlan) === Option(topKThreshold - 1))

val collectLimitStatement = s"SELECT * FROM tv ORDER BY id LIMIT $topKThreshold"
sparkPlan = spark.sql(collectLimitStatement).queryExecution.sparkPlan
assert(sparkPlan.isInstanceOf[CollectLimitExec] || sparkPlan.asInstanceOf[
AdaptiveSparkPlanExec].inputPlan.isInstanceOf[CollectLimitExec])
assert(SparkDatasetHelper.planLimit(sparkPlan) === Option(topKThreshold))
}
}
}

0 comments on commit 367622b

Please sign in to comment.