spark ApplicationInfo 源码
spark ApplicationInfo 代码
文件路径:/core/src/main/scala/org/apache/spark/deploy/master/ApplicationInfo.scala
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.spark.deploy.master
import java.util.Date
import scala.collection.mutable
import scala.collection.mutable.ArrayBuffer
import org.apache.spark.deploy.ApplicationDescription
import org.apache.spark.resource.{ResourceInformation, ResourceProfile, ResourceUtils}
import org.apache.spark.resource.ResourceProfile.DEFAULT_RESOURCE_PROFILE_ID
import org.apache.spark.rpc.RpcEndpointRef
import org.apache.spark.util.Utils
private[spark] class ApplicationInfo(
val startTime: Long,
val id: String,
val desc: ApplicationDescription,
val submitDate: Date,
val driver: RpcEndpointRef,
defaultCores: Int)
extends Serializable {
@transient var state: ApplicationState.Value = _
@transient var executors: mutable.HashMap[Int, ExecutorDesc] = _
@transient var removedExecutors: ArrayBuffer[ExecutorDesc] = _
@transient var coresGranted: Int = _
@transient var endTime: Long = _
@transient var appSource: ApplicationSource = _
@transient private var executorsPerResourceProfileId: mutable.HashMap[Int, mutable.Set[Int]] = _
@transient private var targetNumExecutorsPerResourceProfileId: mutable.HashMap[Int, Int] = _
@transient private var rpIdToResourceProfile: mutable.HashMap[Int, ResourceProfile] = _
@transient private var rpIdToResourceDesc: mutable.HashMap[Int, ExecutorResourceDescription] = _
@transient private var nextExecutorId: Int = _
init()
private def readObject(in: java.io.ObjectInputStream): Unit = Utils.tryOrIOException {
in.defaultReadObject()
init()
}
private def init(): Unit = {
state = ApplicationState.WAITING
executors = new mutable.HashMap[Int, ExecutorDesc]
coresGranted = 0
endTime = -1L
appSource = new ApplicationSource(this)
nextExecutorId = 0
removedExecutors = new ArrayBuffer[ExecutorDesc]
val initialExecutorLimit = desc.initialExecutorLimit.getOrElse(Integer.MAX_VALUE)
rpIdToResourceProfile = new mutable.HashMap[Int, ResourceProfile]()
rpIdToResourceProfile(DEFAULT_RESOURCE_PROFILE_ID) = desc.defaultProfile
rpIdToResourceDesc = new mutable.HashMap[Int, ExecutorResourceDescription]()
createResourceDescForResourceProfile(desc.defaultProfile)
targetNumExecutorsPerResourceProfileId = new mutable.HashMap[Int, Int]()
targetNumExecutorsPerResourceProfileId(DEFAULT_RESOURCE_PROFILE_ID) = initialExecutorLimit
executorsPerResourceProfileId = new mutable.HashMap[Int, mutable.Set[Int]]()
}
private[deploy] def getOrUpdateExecutorsForRPId(rpId: Int): mutable.Set[Int] = {
executorsPerResourceProfileId.getOrElseUpdate(rpId, mutable.HashSet[Int]())
}
private[deploy] def getTargetExecutorNumForRPId(rpId: Int): Int = {
targetNumExecutorsPerResourceProfileId.getOrElse(rpId, 0)
}
private[deploy] def getRequestedRPIds(): Seq[Int] = {
rpIdToResourceProfile.keys.toSeq.sorted
}
private def createResourceDescForResourceProfile(resourceProfile: ResourceProfile): Unit = {
if (!rpIdToResourceDesc.contains(resourceProfile.id)) {
val defaultMemoryMbPerExecutor = desc.memoryPerExecutorMB
val defaultCoresPerExecutor = desc.coresPerExecutor
val coresPerExecutor = resourceProfile.getExecutorCores
.orElse(defaultCoresPerExecutor)
val memoryMbPerExecutor = resourceProfile.getExecutorMemory
.map(_.toInt)
.getOrElse(defaultMemoryMbPerExecutor)
val customResources = ResourceUtils.executorResourceRequestToRequirement(
resourceProfile.getCustomExecutorResources().values.toSeq.sortBy(_.resourceName))
rpIdToResourceDesc(resourceProfile.id) =
ExecutorResourceDescription(coresPerExecutor, memoryMbPerExecutor, customResources)
}
}
// Get resources required for schedule.
private[deploy] def getResourceDescriptionForRpId(rpId: Int): ExecutorResourceDescription = {
rpIdToResourceDesc(rpId)
}
private[deploy] def requestExecutors(
resourceProfileToTotalExecs: Map[ResourceProfile, Int]): Unit = {
resourceProfileToTotalExecs.foreach { case (rp, num) =>
createResourceDescForResourceProfile(rp)
if (!rpIdToResourceProfile.contains(rp.id)) {
rpIdToResourceProfile(rp.id) = rp
}
targetNumExecutorsPerResourceProfileId(rp.id) = num
}
}
private[deploy] def getResourceProfileById(rpId: Int): ResourceProfile = {
rpIdToResourceProfile(rpId)
}
private def newExecutorId(useID: Option[Int] = None): Int = {
useID match {
case Some(id) =>
nextExecutorId = math.max(nextExecutorId, id + 1)
id
case None =>
val id = nextExecutorId
nextExecutorId += 1
id
}
}
private[master] def addExecutor(
worker: WorkerInfo,
cores: Int,
memoryMb: Int,
resources: Map[String, ResourceInformation],
rpId: Int,
useID: Option[Int] = None): ExecutorDesc = {
val exec = new ExecutorDesc(
newExecutorId(useID), this, worker, cores, memoryMb, resources, rpId)
executors(exec.id) = exec
getOrUpdateExecutorsForRPId(rpId).add(exec.id)
coresGranted += cores
exec
}
private[master] def removeExecutor(exec: ExecutorDesc): Unit = {
if (executors.contains(exec.id)) {
removedExecutors += executors(exec.id)
executors -= exec.id
executorsPerResourceProfileId(exec.rpId) -= exec.id
coresGranted -= exec.cores
}
}
private val requestedCores = desc.maxCores.getOrElse(defaultCores)
private[master] def coresLeft: Int = requestedCores - coresGranted
private var _retryCount = 0
private[master] def retryCount = _retryCount
private[master] def incrementRetryCount() = {
_retryCount += 1
_retryCount
}
private[master] def resetRetryCount() = _retryCount = 0
private[master] def markFinished(endState: ApplicationState.Value): Unit = {
state = endState
endTime = System.currentTimeMillis()
}
private[master] def isFinished: Boolean = {
state != ApplicationState.WAITING && state != ApplicationState.RUNNING
}
/**
* Return the total limit on the number of executors for all resource profiles.
*/
private[deploy] def getExecutorLimit: Int = {
targetNumExecutorsPerResourceProfileId.values.sum
}
def duration: Long = {
if (endTime != -1) {
endTime - startTime
} else {
System.currentTimeMillis() - startTime
}
}
}
相关信息
相关文章
spark ExecutorResourceDescription 源码
0
赞
- 所属分类: 前端技术
- 本文标签:
热门推荐
-
2、 - 优质文章
-
3、 gate.io
-
8、 golang
-
9、 openharmony
-
10、 Vue中input框自动聚焦