diff --git a/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala b/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
index 6cb43c02b8f08..2d8c3b949c1ac 100644
--- a/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
+++ b/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
@@ -79,6 +79,7 @@ private[ui] class ExecutorsPage(parent: ExecutorsTab) extends WebUIPage("") {
val maximumMemory = values("Maximum Memory")
val memoryUsed = values("Memory Used")
val diskUsed = values("Disk Used")
+ // scalastyle:off
{values("Executor ID")} |
{values("Address")} |
@@ -94,10 +95,11 @@ private[ui] class ExecutorsPage(parent: ExecutorsTab) extends WebUIPage("") {
{values("Failed Tasks")} |
{values("Complete Tasks")} |
{values("Total Tasks")} |
- {Utils.msDurationToString(values("Task Time").toLong)} |
- {Utils.bytesToString(values("Shuffle Read").toLong)} |
- {Utils.bytesToString(values("Shuffle Write").toLong)} |
+ {Utils.msDurationToString(values("Task Time").toLong)} |
+ {Utils.bytesToString(values("Shuffle Read").toLong)} |
+ {Utils.bytesToString(values("Shuffle Write").toLong)} |
+ // scalastyle:on
}
/** Represent an executor's info as a map given a storage status index */
diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/ExecutorTable.scala b/core/src/main/scala/org/apache/spark/ui/jobs/ExecutorTable.scala
index c83e196c9c156..add0e9878a546 100644
--- a/core/src/main/scala/org/apache/spark/ui/jobs/ExecutorTable.scala
+++ b/core/src/main/scala/org/apache/spark/ui/jobs/ExecutorTable.scala
@@ -67,18 +67,20 @@ private[ui] class ExecutorTable(stageId: Int, parent: JobProgressTab) {
executorIdToSummary match {
case Some(x) =>
x.toSeq.sortBy(_._1).map { case (k, v) => {
+ // scalastyle:off
{k} |
{executorIdToAddress.getOrElse(k, "CANNOT FIND ADDRESS")} |
- {UIUtils.formatDuration(v.taskTime)} |
+ {UIUtils.formatDuration(v.taskTime)} |
{v.failedTasks + v.succeededTasks} |
{v.failedTasks} |
{v.succeededTasks} |
- {Utils.bytesToString(v.shuffleRead)} |
- {Utils.bytesToString(v.shuffleWrite)} |
- {Utils.bytesToString(v.memoryBytesSpilled)} |
- {Utils.bytesToString(v.diskBytesSpilled)} |
+ {Utils.bytesToString(v.shuffleRead)} |
+ {Utils.bytesToString(v.shuffleWrite)} |
+ {Utils.bytesToString(v.memoryBytesSpilled)} |
+ {Utils.bytesToString(v.diskBytesSpilled)} |
+ // scalastyle:on
}
}
case _ => Seq[Node]()
diff --git a/core/src/main/scala/org/apache/spark/ui/storage/StoragePage.scala b/core/src/main/scala/org/apache/spark/ui/storage/StoragePage.scala
index b66edd91f56c0..9813d9330ac7f 100644
--- a/core/src/main/scala/org/apache/spark/ui/storage/StoragePage.scala
+++ b/core/src/main/scala/org/apache/spark/ui/storage/StoragePage.scala
@@ -49,6 +49,7 @@ private[ui] class StoragePage(parent: StorageTab) extends WebUIPage("") {
/** Render an HTML row representing an RDD */
private def rddRow(rdd: RDDInfo): Seq[Node] = {
+ // scalastyle:off
@@ -59,9 +60,10 @@ private[ui] class StoragePage(parent: StorageTab) extends WebUIPage("") {
|
{rdd.numCachedPartitions} |
{"%.0f%%".format(rdd.numCachedPartitions * 100.0 / rdd.numPartitions)} |
- {Utils.bytesToString(rdd.memSize)} |
- {Utils.bytesToString(rdd.tachyonSize)} |
- {Utils.bytesToString(rdd.diskSize)} |
+ {Utils.bytesToString(rdd.memSize)} |
+ {Utils.bytesToString(rdd.tachyonSize)} |
+ {Utils.bytesToString(rdd.diskSize)} |
+ // scalastyle:on
}
}