From 3ff698b12b5a6d5edecb223ae8df71042de7eb9c Mon Sep 17 00:00:00 2001 From: Michael Munday Date: Tue, 30 Jun 2020 15:06:13 +0100 Subject: [PATCH] [SPARK-32458][SQL][TESTS] Fix incorrectly sized row value reads. Use getInt to fetch 32 bit values and getLong to fetch 64 bit values. The mismatched sizes caused test failures on big-endian systems. --- .../spark/sql/catalyst/encoders/RowEncoderSuite.scala | 2 +- .../apache/spark/sql/catalyst/util/UnsafeMapSuite.scala | 8 ++++---- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala index fd24f058f357c..d20a9ba3f0f68 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/encoders/RowEncoderSuite.scala @@ -336,7 +336,7 @@ class RowEncoderSuite extends CodegenInterpretedPlanTest { val encoder = RowEncoder(schema).resolveAndBind() val localDate = java.time.LocalDate.parse("2019-02-27") val row = toRow(encoder, Row(localDate)) - assert(row.getLong(0) === DateTimeUtils.localDateToDays(localDate)) + assert(row.getInt(0) === DateTimeUtils.localDateToDays(localDate)) val readback = fromRow(encoder, row) assert(readback.get(0).equals(localDate)) } diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeMapSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeMapSuite.scala index ebc88612be22a..443534fd0a06a 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeMapSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/UnsafeMapSuite.scala @@ -48,8 +48,8 @@ class UnsafeMapSuite extends SparkFunSuite { val ser = new JavaSerializer(new SparkConf).newInstance() val mapDataSer = ser.deserialize[UnsafeMapData](ser.serialize(unsafeMapData)) assert(mapDataSer.numElements() == 1) - assert(mapDataSer.keyArray().getInt(0) == 19285) - assert(mapDataSer.valueArray().getInt(0) == 19286) + assert(mapDataSer.keyArray().getLong(0) == 19285) + assert(mapDataSer.valueArray().getLong(0) == 19286) assert(mapDataSer.getBaseObject.asInstanceOf[Array[Byte]].length == 1024) } @@ -57,8 +57,8 @@ class UnsafeMapSuite extends SparkFunSuite { val ser = new KryoSerializer(new SparkConf).newInstance() val mapDataSer = ser.deserialize[UnsafeMapData](ser.serialize(unsafeMapData)) assert(mapDataSer.numElements() == 1) - assert(mapDataSer.keyArray().getInt(0) == 19285) - assert(mapDataSer.valueArray().getInt(0) == 19286) + assert(mapDataSer.keyArray().getLong(0) == 19285) + assert(mapDataSer.valueArray().getLong(0) == 19286) assert(mapDataSer.getBaseObject.asInstanceOf[Array[Byte]].length == 1024) } }