diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/InsertIntoHadoopFsRelationCommand.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/InsertIntoHadoopFsRelationCommand.scala index 8e73c10bc0f5b..fe6ec094812e8 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/InsertIntoHadoopFsRelationCommand.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/InsertIntoHadoopFsRelationCommand.scala @@ -174,7 +174,6 @@ case class InsertIntoHadoopFsRelationCommand( qualifiedOutputPath } - val statsTrackers = Seq(basicWriteJobStatsTracker(hadoopConf)) val updatedPartitionPaths = FileFormatWriter.write( sparkSession = sparkSession, @@ -186,10 +185,11 @@ case class InsertIntoHadoopFsRelationCommand( hadoopConf = hadoopConf, partitionColumns = partitionColumns, bucketSpec = bucketSpec, - statsTrackers = statsTrackers, + statsTrackers = Seq(basicWriteJobStatsTracker(hadoopConf)), options = options, numStaticPartitionCols = staticPartitions.size) + // update metastore partition metadata if (updatedPartitionPaths.isEmpty && staticPartitions.nonEmpty && partitionColumns.length == staticPartitions.size) {