-
Notifications
You must be signed in to change notification settings - Fork 29k
[SPARK-27676][SQL][SS] InMemoryFileIndex should respect spark.sql.files.ignoreMissingFiles #24668
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
Closed
Changes from 4 commits
Commits
Show all changes
13 commits
Select commit
Hold shift + click to select a range
05f9228
Only ignore FileNotFoundException when spark.sql.files.ignoreMissingF…
joshrosen-stripe a31c08a
Update test cases to reflect behavior change
JoshRosen 69f8db6
Only non-root deletions should respect flag.
JoshRosen 24ad834
Remove debug code
JoshRosen 400a02b
Fix indentation
JoshRosen 0c1eba3
Add note to migration guide
JoshRosen 88dc6b6
Test with parallel partition discovery
JoshRosen 86c3a9d
Clarify migration guide comment
JoshRosen 42e8b98
Strengthen test assertions further (to fix bug in tests and guard aga…
JoshRosen 2a6240b
Work in progress towards fixing races for root file deletion
JoshRosen 97bac91
Revert "Work in progress towards fixing races for root file deletion"
JoshRosen 58e9544
Update comment to clarify exception for root paths
JoshRosen d9c5903
Merge remote-tracking branch 'origin/master' into SPARK-27676
JoshRosen File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There is one config
parallelPartitionDiscoveryThresholdcan control code path of partition discovery. With the default value, this only tests serial listing?There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good point. In the case of a parallel listing, this would cause the listing Spark job to fail with a
FileNotFoundException(after maxTaskRetries attempts to list the missing file).In the interests of complete test coverage, I'll update the test case to exercise the parallel listing path, too.
(Combinatorial test coverage is hard!)