diff --git a/core/src/main/scala/org/apache/spark/CacheManager.scala b/core/src/main/scala/org/apache/spark/CacheManager.scala index 6ea0f1404e951..c3a079d64db7a 100644 --- a/core/src/main/scala/org/apache/spark/CacheManager.scala +++ b/core/src/main/scala/org/apache/spark/CacheManager.scala @@ -155,7 +155,7 @@ private[spark] class CacheManager(blockManager: BlockManager) extends Logging { * single partition. Instead, we unfold the values cautiously, potentially aborting and * dropping the partition to disk if applicable. */ - blockManager.memoryStore.unfoldSafely(key, values, storageLevel, updatedBlocks) match { + blockManager.memoryStore.unfoldSafely(key, values, updatedBlocks) match { case Left(arrayValues) => // We have successfully unfolded the entire partition, so cache it in memory updatedBlocks ++= blockManager.put(key, arrayValues, storageLevel, tellMaster = true) diff --git a/core/src/main/scala/org/apache/spark/storage/MemoryStore.scala b/core/src/main/scala/org/apache/spark/storage/MemoryStore.scala index 78d16a4863567..c102f5b36c2ad 100644 --- a/core/src/main/scala/org/apache/spark/storage/MemoryStore.scala +++ b/core/src/main/scala/org/apache/spark/storage/MemoryStore.scala @@ -98,7 +98,7 @@ private[spark] class MemoryStore(blockManager: BlockManager, maxMemory: Long) level: StorageLevel, returnValues: Boolean): PutResult = { val droppedBlocks = new ArrayBuffer[(BlockId, BlockStatus)] - val unfoldedValues = unfoldSafely(blockId, values, level, droppedBlocks) + val unfoldedValues = unfoldSafely(blockId, values, droppedBlocks) unfoldedValues match { case Left(arrayValues) => // Values are fully unfolded in memory, so store them as an array @@ -187,7 +187,6 @@ private[spark] class MemoryStore(blockManager: BlockManager, maxMemory: Long) def unfoldSafely( blockId: BlockId, values: Iterator[Any], - storageLevel: StorageLevel, droppedBlocks: ArrayBuffer[(BlockId, BlockStatus)]) : Either[Array[Any], Iterator[Any]] = {