Skip to content

Conversation

@dgshep
Copy link
Contributor

@dgshep dgshep commented Apr 6, 2014

While working on spark-1337 with @pwendell, we noticed that not all of the metadata maps in JobProgessListener were being properly cleaned. This could lead to a (hypothetical) memory leak issue should a job run long enough. This patch aims to address the issue.

@AmplabJenkins
Copy link

Can one of the admins verify this patch?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is there a reason why you removed this check? If you look where stageIdToDescription is populated, it may or may contain an an entry for the stage.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ah I see - this is because we're using the Scala HashMap?

scala> import scala.collection.mutable.HashMap
import scala.collection.mutable.HashMap
scala> val x = new HashMap[String, String]()
x: scala.collection.mutable.HashMap[String,String] = Map()
scala> x.remove("nonexistent")
res0: Option[String] = None

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yup! No need for that check. :)

@pwendell
Copy link
Contributor

pwendell commented Apr 6, 2014

Okay this looks good to me. Do you mind creating a new JIRA for this and setting the title to something like: "stageIdToExecutorSummaries can lead to memory leak". And could you set the title of this to reflect the JIRA number?

This is a bit of extra overhead, but may run into this leak and it would be good to have a reference for it.

@dgshep
Copy link
Contributor Author

dgshep commented Apr 7, 2014

@dgshep dgshep changed the title In the contex of SPARK-1337: Make sure that all metadata fields are properly cleaned SPARK-1432: Make sure that all metadata fields are properly cleaned Apr 7, 2014
@pwendell
Copy link
Contributor

pwendell commented Apr 7, 2014

Thanks - merged this into master and 0.9 branch.

@asfgit asfgit closed this in a3c51c6 Apr 7, 2014
asfgit pushed a commit that referenced this pull request Apr 7, 2014
While working on spark-1337 with @pwendell, we noticed that not all of the metadata maps in JobProgessListener were being properly cleaned. This could lead to a (hypothetical) memory leak issue should a job run long enough. This patch aims to address the issue.

Author: Davis Shepherd <[email protected]>

Closes #338 from dgshep/master and squashes the following commits:

a77b65c [Davis Shepherd] In the contex of SPARK-1337: Make sure that all metadata fields are properly cleaned
(cherry picked from commit a3c51c6)

Conflicts:

	core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala
andrewor14 pushed a commit to andrewor14/spark that referenced this pull request Apr 7, 2014
pdeyhim pushed a commit to pdeyhim/spark-1 that referenced this pull request Jun 25, 2014
While working on spark-1337 with @pwendell, we noticed that not all of the metadata maps in JobProgessListener were being properly cleaned. This could lead to a (hypothetical) memory leak issue should a job run long enough. This patch aims to address the issue.

Author: Davis Shepherd <[email protected]>

Closes apache#338 from dgshep/master and squashes the following commits:

a77b65c [Davis Shepherd] In the contex of SPARK-1337: Make sure that all metadata fields are properly cleaned
mccheah pushed a commit to mccheah/spark that referenced this pull request Oct 3, 2018
bzhaoopenstack pushed a commit to bzhaoopenstack/spark that referenced this pull request Sep 11, 2019
Run cluster-api-provider-openstack-image-build in sigs.k8s.io
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants