-
Notifications
You must be signed in to change notification settings - Fork 29k
SPARK-1432: Make sure that all metadata fields are properly cleaned #338
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Can one of the admins verify this patch? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is there a reason why you removed this check? If you look where stageIdToDescription is populated, it may or may contain an an entry for the stage.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah I see - this is because we're using the Scala HashMap?
scala> import scala.collection.mutable.HashMap
import scala.collection.mutable.HashMap
scala> val x = new HashMap[String, String]()
x: scala.collection.mutable.HashMap[String,String] = Map()
scala> x.remove("nonexistent")
res0: Option[String] = None
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yup! No need for that check. :)
|
Okay this looks good to me. Do you mind creating a new JIRA for this and setting the title to something like: "stageIdToExecutorSummaries can lead to memory leak". And could you set the title of this to reflect the JIRA number? This is a bit of extra overhead, but may run into this leak and it would be good to have a reference for it. |
|
Thanks - merged this into master and 0.9 branch. |
While working on spark-1337 with @pwendell, we noticed that not all of the metadata maps in JobProgessListener were being properly cleaned. This could lead to a (hypothetical) memory leak issue should a job run long enough. This patch aims to address the issue. Author: Davis Shepherd <[email protected]> Closes #338 from dgshep/master and squashes the following commits: a77b65c [Davis Shepherd] In the contex of SPARK-1337: Make sure that all metadata fields are properly cleaned (cherry picked from commit a3c51c6) Conflicts: core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala
SPARK-1005 Ning upgrade
While working on spark-1337 with @pwendell, we noticed that not all of the metadata maps in JobProgessListener were being properly cleaned. This could lead to a (hypothetical) memory leak issue should a job run long enough. This patch aims to address the issue. Author: Davis Shepherd <[email protected]> Closes apache#338 from dgshep/master and squashes the following commits: a77b65c [Davis Shepherd] In the contex of SPARK-1337: Make sure that all metadata fields are properly cleaned
Run cluster-api-provider-openstack-image-build in sigs.k8s.io
While working on spark-1337 with @pwendell, we noticed that not all of the metadata maps in JobProgessListener were being properly cleaned. This could lead to a (hypothetical) memory leak issue should a job run long enough. This patch aims to address the issue.