Skip to content

Commit

Permalink
Update PySpark to be less OOM-prone as well
Browse files Browse the repository at this point in the history
  • Loading branch information
ash211 committed Sep 3, 2014
1 parent 3a156b8 commit 09a27f7
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion python/pyspark/rdd.py
Original file line number Diff line number Diff line change
Expand Up @@ -1035,7 +1035,7 @@ def take(self, num):
# try all partitions next. Otherwise, interpolate the number
# of partitions we need to try, but overestimate it by 50%.
if len(items) == 0:
numPartsToTry = totalParts - 1
numPartsToTry = partsScanned * 2
else:
numPartsToTry = int(1.5 * num * partsScanned / len(items))

Expand Down

0 comments on commit 09a27f7

Please sign in to comment.