Skip to content

Commit

Permalink
Merge remote-tracking branch 'upstream/master' into writeSupportFollowup
Browse files Browse the repository at this point in the history
Conflicts:
	python/pyspark/sql.py
	python/pyspark/sql/tests.py
  • Loading branch information
yhuai committed Feb 10, 2015
2 parents 0832ce4 + 31d435e commit 3abc215
Show file tree
Hide file tree
Showing 23 changed files with 3,208 additions and 2,874 deletions.
3 changes: 3 additions & 0 deletions core/src/main/scala/org/apache/spark/SparkContext.scala
Original file line number Diff line number Diff line change
Expand Up @@ -1420,6 +1420,9 @@ class SparkContext(config: SparkConf) extends Logging with ExecutorAllocationCli
val callSite = getCallSite
val cleanedFunc = clean(func)
logInfo("Starting job: " + callSite.shortForm)
if (conf.getBoolean("spark.logLineage", false)) {
logInfo("RDD's recursive dependencies:\n" + rdd.toDebugString)
}
dagScheduler.runJob(rdd, cleanedFunc, partitions, callSite, allowLocal,
resultHandler, localProperties.get)
progressBar.foreach(_.finishAll())
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -314,6 +314,11 @@ class CoarseGrainedSchedulerBackend(scheduler: TaskSchedulerImpl, val actorSyste
* Return whether the request is acknowledged.
*/
final override def requestExecutors(numAdditionalExecutors: Int): Boolean = synchronized {
if (numAdditionalExecutors < 0) {
throw new IllegalArgumentException(
"Attempted to request a negative number of additional executor(s) " +
s"$numAdditionalExecutors from the cluster manager. Please specify a positive number!")
}
logInfo(s"Requesting $numAdditionalExecutors additional executor(s) from the cluster manager")
logDebug(s"Number of pending executors is now $numPendingExecutors")
numPendingExecutors += numAdditionalExecutors
Expand Down
6 changes: 3 additions & 3 deletions dev/run-tests
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ function handle_error () {
}


# Build against the right verison of Hadoop.
# Build against the right version of Hadoop.
{
if [ -n "$AMPLAB_JENKINS_BUILD_PROFILE" ]; then
if [ "$AMPLAB_JENKINS_BUILD_PROFILE" = "hadoop1.0" ]; then
Expand Down Expand Up @@ -77,7 +77,7 @@ export SBT_MAVEN_PROFILES_ARGS="$SBT_MAVEN_PROFILES_ARGS -Pkinesis-asl"
fi
}

# Only run Hive tests if there are sql changes.
# Only run Hive tests if there are SQL changes.
# Partial solution for SPARK-1455.
if [ -n "$AMPLAB_JENKINS" ]; then
git fetch origin master:master
Expand Down Expand Up @@ -183,7 +183,7 @@ CURRENT_BLOCK=$BLOCK_SPARK_UNIT_TESTS
if [ -n "$_SQL_TESTS_ONLY" ]; then
# This must be an array of individual arguments. Otherwise, having one long string
# will be interpreted as a single test, which doesn't work.
SBT_MAVEN_TEST_ARGS=("catalyst/test" "sql/test" "hive/test" "mllib/test")
SBT_MAVEN_TEST_ARGS=("catalyst/test" "sql/test" "hive/test" "hive-thriftserver/test" "mllib/test")
else
SBT_MAVEN_TEST_ARGS=("test")
fi
Expand Down
7 changes: 2 additions & 5 deletions python/docs/pyspark.ml.rst
Original file line number Diff line number Diff line change
@@ -1,11 +1,8 @@
pyspark.ml package
=====================

Submodules
----------

pyspark.ml module
-----------------
Module Context
--------------

.. automodule:: pyspark.ml
:members:
Expand Down
3 changes: 0 additions & 3 deletions python/docs/pyspark.mllib.rst
Original file line number Diff line number Diff line change
@@ -1,9 +1,6 @@
pyspark.mllib package
=====================

Submodules
----------

pyspark.mllib.classification module
-----------------------------------

Expand Down
12 changes: 10 additions & 2 deletions python/docs/pyspark.sql.rst
Original file line number Diff line number Diff line change
@@ -1,10 +1,18 @@
pyspark.sql module
==================

Module contents
---------------
Module Context
--------------

.. automodule:: pyspark.sql
:members:
:undoc-members:
:show-inheritance:


pyspark.sql.types module
------------------------
.. automodule:: pyspark.sql.types
:members:
:undoc-members:
:show-inheritance:
2 changes: 1 addition & 1 deletion python/pyspark/mllib/linalg.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@

import numpy as np

from pyspark.sql import UserDefinedType, StructField, StructType, ArrayType, DoubleType, \
from pyspark.sql.types import UserDefinedType, StructField, StructType, ArrayType, DoubleType, \
IntegerType, ByteType


Expand Down
Loading

0 comments on commit 3abc215

Please sign in to comment.