diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/TransformExpression.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/TransformExpression.scala index 98b5c641096fb..c388e7691dd58 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/TransformExpression.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/TransformExpression.scala @@ -133,15 +133,6 @@ case class TransformExpression( } } - /** - * Returns Java source code that can be compiled to evaluate this expression. - * The default behavior is to call the eval method of the expression. Concrete expression - * implementations should override this to do actual code generation. - * - * @param ctx a [[CodegenContext]] - * @param ev an [[ExprCode]] with unique terms. - * @return an [[ExprCode]] containing the Java source code to generate the given expression - */ override protected def doGenCode(ctx: CodegenContext, ev: ExprCode): ExprCode = throw QueryExecutionErrors.cannotGenerateCodeForExpressionError(this) } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/connector/WriteDistributionAndOrderingSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/connector/WriteDistributionAndOrderingSuite.scala index e301cd8e3f359..4d9c001aebb13 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/connector/WriteDistributionAndOrderingSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/connector/WriteDistributionAndOrderingSuite.scala @@ -20,7 +20,8 @@ package org.apache.spark.sql.connector import java.sql.Date import java.util.Collections -import org.apache.spark.sql.{AnalysisException, DataFrame, Row, catalyst} + +import org.apache.spark.sql.{catalyst, AnalysisException, DataFrame, Row} import org.apache.spark.sql.catalyst.expressions.{ApplyFunctionExpression, Cast, Literal} import org.apache.spark.sql.catalyst.expressions.objects.StaticInvoke import org.apache.spark.sql.catalyst.plans.physical