From 55e27af45f4a39bb1e86425be5687af3608fd4f3 Mon Sep 17 00:00:00 2001 From: Josh Rosen Date: Wed, 3 Feb 2016 16:02:49 -0800 Subject: [PATCH] Remove assertions which guarded against misplanning. --- .../main/scala/org/apache/spark/sql/execution/limit.scala | 8 -------- 1 file changed, 8 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/limit.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/limit.scala index f72fa972fd5cf..5acde5996b543 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/limit.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/limit.scala @@ -44,14 +44,6 @@ trait BaseLimit extends UnaryNode { override def output: Seq[Attribute] = child.output override def outputOrdering: Seq[SortOrder] = child.outputOrdering override def outputPartitioning: Partitioning = child.outputPartitioning - override def executeTake(n: Int): Array[InternalRow] = { - throw new UnsupportedOperationException( - s"Should not invoke executeTake() on ${getClass.getName}; use CollectLimit instead.") - } - override def executeCollect(): Array[InternalRow] = { - throw new UnsupportedOperationException( - s"Should not invoke executeCollect() on ${getClass.getName}; use CollectLimit instead.") - } protected override def doExecute(): RDD[InternalRow] = child.execute().mapPartitions { iter => iter.take(limit) }