diff --git a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala index 9ed49e01be639524575b0747a169ace1f062ec9b..95bd62e88db2b718fa50f0ed64177e2ea4de81b8 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/Master.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/Master.scala @@ -51,7 +51,7 @@ private[spark] class Master( val conf = new SparkConf - val DATE_FORMAT = new SimpleDateFormat("yyyyMMddHHmmss") // For application IDs + def createDateFormat = new SimpleDateFormat("yyyyMMddHHmmss") // For application IDs val WORKER_TIMEOUT = conf.getLong("spark.worker.timeout", 60) * 1000 val RETAINED_APPLICATIONS = conf.getInt("spark.deploy.retainedApplications", 200) val REAPER_ITERATIONS = conf.getInt("spark.dead.worker.persistence", 15) @@ -682,7 +682,7 @@ private[spark] class Master( /** Generate a new app ID given a app's submission date */ def newApplicationId(submitDate: Date): String = { - val appId = "app-%s-%04d".format(DATE_FORMAT.format(submitDate), nextAppNumber) + val appId = "app-%s-%04d".format(createDateFormat.format(submitDate), nextAppNumber) nextAppNumber += 1 appId } @@ -706,7 +706,7 @@ private[spark] class Master( } def newDriverId(submitDate: Date): String = { - val appId = "driver-%s-%04d".format(DATE_FORMAT.format(submitDate), nextDriverNumber) + val appId = "driver-%s-%04d".format(createDateFormat.format(submitDate), nextDriverNumber) nextDriverNumber += 1 appId } diff --git a/core/src/main/scala/org/apache/spark/deploy/master/ui/IndexPage.scala b/core/src/main/scala/org/apache/spark/deploy/master/ui/IndexPage.scala index 7ec71eb80bfc07fd4f0a17e11813c7410dda2687..8c1d6c7cce45057b65e1d1abb581343d15055e4e 100644 --- a/core/src/main/scala/org/apache/spark/deploy/master/ui/IndexPage.scala +++ b/core/src/main/scala/org/apache/spark/deploy/master/ui/IndexPage.scala @@ -25,10 +25,10 @@ import scala.xml.Node import akka.pattern.ask import org.json4s.JValue -import org.apache.spark.deploy.{DeployWebUI, JsonProtocol} +import org.apache.spark.deploy.{JsonProtocol} import org.apache.spark.deploy.DeployMessages.{MasterStateResponse, RequestMasterState} import org.apache.spark.deploy.master.{ApplicationInfo, DriverInfo, WorkerInfo} -import org.apache.spark.ui.UIUtils +import org.apache.spark.ui.{WebUI, UIUtils} import org.apache.spark.util.Utils private[spark] class IndexPage(parent: MasterWebUI) { @@ -169,10 +169,10 @@ private[spark] class IndexPage(parent: MasterWebUI) { <td sorttable_customkey={app.desc.memoryPerSlave.toString}> {Utils.megabytesToString(app.desc.memoryPerSlave)} </td> - <td>{DeployWebUI.formatDate(app.submitDate)}</td> + <td>{WebUI.formatDate(app.submitDate)}</td> <td>{app.desc.user}</td> <td>{app.state.toString}</td> - <td>{DeployWebUI.formatDuration(app.duration)}</td> + <td>{WebUI.formatDuration(app.duration)}</td> </tr> } diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala index 5e0fc31fff22f57f5273c0e7868e1068f54aefc0..8a71ddda4cb5efeb4fb53b3320283e35b17cb5c6 100755 --- a/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/Worker.scala @@ -56,7 +56,7 @@ private[spark] class Worker( Utils.checkHost(host, "Expected hostname") assert (port > 0) - val DATE_FORMAT = new SimpleDateFormat("yyyyMMddHHmmss") // For worker and executor IDs + def createDateFormat = new SimpleDateFormat("yyyyMMddHHmmss") // For worker and executor IDs // Send a heartbeat every (heartbeat timeout) / 4 milliseconds val HEARTBEAT_MILLIS = conf.getLong("spark.worker.timeout", 60) * 1000 / 4 @@ -319,7 +319,7 @@ private[spark] class Worker( } def generateWorkerId(): String = { - "worker-%s-%s-%d".format(DATE_FORMAT.format(new Date), host, port) + "worker-%s-%s-%d".format(createDateFormat.format(new Date), host, port) } override def postStop() { diff --git a/core/src/main/scala/org/apache/spark/scheduler/JobLogger.scala b/core/src/main/scala/org/apache/spark/scheduler/JobLogger.scala index b3a67d7e179763d4c0b731844659b2423c9f6d19..5cecf9416b32c680fbf46e2a9b19c953f4266cb2 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/JobLogger.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/JobLogger.scala @@ -55,7 +55,9 @@ class JobLogger(val user: String, val logDirName: String) extends SparkListener private val jobIdToPrintWriter = new HashMap[Int, PrintWriter] private val stageIdToJobId = new HashMap[Int, Int] private val jobIdToStageIds = new HashMap[Int, Seq[Int]] - private val DATE_FORMAT = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") + private val dateFormat = new ThreadLocal[SimpleDateFormat]() { + override def initialValue() = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") + } private val eventQueue = new LinkedBlockingQueue[SparkListenerEvent] createLogDir() @@ -128,7 +130,7 @@ class JobLogger(val user: String, val logDirName: String) extends SparkListener var writeInfo = info if (withTime) { val date = new Date(System.currentTimeMillis()) - writeInfo = DATE_FORMAT.format(date) + ": " + info + writeInfo = dateFormat.get.format(date) + ": " + info } jobIdToPrintWriter.get(jobId).foreach(_.println(writeInfo)) } diff --git a/core/src/main/scala/org/apache/spark/deploy/WebUI.scala b/core/src/main/scala/org/apache/spark/ui/WebUI.scala similarity index 74% rename from core/src/main/scala/org/apache/spark/deploy/WebUI.scala rename to core/src/main/scala/org/apache/spark/ui/WebUI.scala index ae258b58b9cc5554ecc883ade04f9aa87fc98c5e..a7b872f3445a4fc8d2ec1af3938ef7fef3db005b 100644 --- a/core/src/main/scala/org/apache/spark/deploy/WebUI.scala +++ b/core/src/main/scala/org/apache/spark/ui/WebUI.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.deploy +package org.apache.spark.ui import java.text.SimpleDateFormat import java.util.Date @@ -23,12 +23,15 @@ import java.util.Date /** * Utilities used throughout the web UI. */ -private[spark] object DeployWebUI { - val DATE_FORMAT = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") +private[spark] object WebUI { + // SimpleDateFormat is not thread-safe. Don't expose it to avoid improper use. + private val dateFormat = new ThreadLocal[SimpleDateFormat]() { + override def initialValue(): SimpleDateFormat = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") + } - def formatDate(date: Date): String = DATE_FORMAT.format(date) + def formatDate(date: Date): String = dateFormat.get.format(date) - def formatDate(timestamp: Long): String = DATE_FORMAT.format(new Date(timestamp)) + def formatDate(timestamp: Long): String = dateFormat.get.format(new Date(timestamp)) def formatDuration(milliseconds: Long): String = { val seconds = milliseconds.toDouble / 1000 diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressUI.scala b/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressUI.scala index ee4e9c69c1bd1df5efd1458cd4b8e0aed4a1c8d9..b2c67381cc3daf8208902563ee067ace6a976eb6 100644 --- a/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressUI.scala +++ b/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressUI.scala @@ -17,7 +17,6 @@ package org.apache.spark.ui.jobs -import java.text.SimpleDateFormat import javax.servlet.http.HttpServletRequest import org.eclipse.jetty.servlet.ServletContextHandler @@ -32,7 +31,6 @@ import org.apache.spark.util.Utils private[ui] class JobProgressUI(parent: SparkUI) { val appName = parent.appName val basePath = parent.basePath - val dateFmt = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") val live = parent.live val sc = parent.sc diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala b/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala index da7f20233063eca2aa74e66cb34d285162698ce6..0c55f2ee7e944cb294f610d2a1223594dea2062d 100644 --- a/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala +++ b/core/src/main/scala/org/apache/spark/ui/jobs/StagePage.scala @@ -23,14 +23,13 @@ import javax.servlet.http.HttpServletRequest import scala.xml.Node import org.apache.spark.ui.Page._ -import org.apache.spark.ui.UIUtils +import org.apache.spark.ui.{WebUI, UIUtils} import org.apache.spark.util.{Utils, Distribution} /** Page showing statistics and task list for a given stage */ private[ui] class StagePage(parent: JobProgressUI) { private val appName = parent.appName private val basePath = parent.basePath - private val dateFmt = parent.dateFmt private lazy val listener = parent.listener def render(request: HttpServletRequest): Seq[Node] = { @@ -253,7 +252,7 @@ private[ui] class StagePage(parent: JobProgressUI) { <td>{info.status}</td> <td>{info.taskLocality}</td> <td>{info.host}</td> - <td>{dateFmt.format(new Date(info.launchTime))}</td> + <td>{WebUI.formatDate(new Date(info.launchTime))}</td> <td sorttable_customkey={duration.toString}> {formatDuration} </td> diff --git a/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala b/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala index 68fef5234c9abe2609ded67820584a2ad4ff07fe..5bf1c95cd5530620c795c8b6f79dc76e51685e3c 100644 --- a/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala +++ b/core/src/main/scala/org/apache/spark/ui/jobs/StageTable.scala @@ -23,13 +23,12 @@ import scala.collection.mutable.HashMap import scala.xml.Node import org.apache.spark.scheduler.{StageInfo, TaskInfo} -import org.apache.spark.ui.UIUtils +import org.apache.spark.ui.{WebUI, UIUtils} import org.apache.spark.util.Utils /** Page showing list of all ongoing and recently finished stages */ private[ui] class StageTable(stages: Seq[StageInfo], parent: JobProgressUI) { private val basePath = parent.basePath - private val dateFmt = parent.dateFmt private lazy val listener = parent.listener private lazy val isFairScheduler = parent.isFairScheduler @@ -82,7 +81,7 @@ private[ui] class StageTable(stages: Seq[StageInfo], parent: JobProgressUI) { val description = listener.stageIdToDescription.get(s.stageId) .map(d => <div><em>{d}</em></div><div>{nameLink}</div>).getOrElse(nameLink) val submissionTime = s.submissionTime match { - case Some(t) => dateFmt.format(new Date(t)) + case Some(t) => WebUI.formatDate(new Date(t)) case None => "Unknown" } val finishTime = s.completionTime.getOrElse(System.currentTimeMillis) diff --git a/core/src/main/scala/org/apache/spark/util/FileLogger.scala b/core/src/main/scala/org/apache/spark/util/FileLogger.scala index f07962096a32c36f6b114a8beac8c3f9c4cfef56..a0c07e32fdc9877942c5343285a7cb0790fc24aa 100644 --- a/core/src/main/scala/org/apache/spark/util/FileLogger.scala +++ b/core/src/main/scala/org/apache/spark/util/FileLogger.scala @@ -44,7 +44,10 @@ class FileLogger( overwrite: Boolean = true) extends Logging { - private val DATE_FORMAT = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") + private val dateFormat = new ThreadLocal[SimpleDateFormat]() { + override def initialValue(): SimpleDateFormat = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss") + } + private val fileSystem = Utils.getHadoopFileSystem(new URI(logDir)) private var fileIndex = 0 @@ -111,7 +114,7 @@ class FileLogger( def log(msg: String, withTime: Boolean = false) { val writeInfo = if (!withTime) msg else { val date = new Date(System.currentTimeMillis()) - DATE_FORMAT.format(date) + ": " + msg + dateFormat.get.format(date) + ": " + msg } writer.foreach(_.print(writeInfo)) }