-
Notifications
You must be signed in to change notification settings - Fork 702
[CARBONDATA-2989] Upgrade spark integration version to 2.3.2 #2779
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
| classOf[Seq[Attribute]], | ||
| classOf[SparkPlan]) | ||
| method.invoke(dataSourceObj, mode, query, query.output, physicalPlan) | ||
| method.invoke(dataSourceObj, mode, query, query.output.map(_.name), physicalPlan) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The parameters of 'writeAndRead' method had been changed, please see: SPARK-PR#22346
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
added some comments too
| override lazy val (outputPartitioning, outputOrdering): (Partitioning, Seq[SortOrder]) = | ||
| (partitioning, Nil) | ||
|
|
||
| override lazy val metadata: Map[String, String] = md |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The parameters (supportsBatch, outputPartitioning, outputOrdering, metadata) had been added keyword 'lazy', please see: SPARK-PR#21815
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: made lazy since spark 2.3.2 version (SPARK-PR#21815)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
fixed
|
@jackylk @chenliang613 @sujith71955 please review. |
|
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/615/ |
|
Thanks for raising the PR, It will better if you can add the description about the changes in this PR. |
|
Build Failed with Spark 2.3.1, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/8876/ |
|
Build Success with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/807/ |
|
retest this please |
| @@ -0,0 +1,55 @@ | |||
| /* | |||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why need to move CarbonDataSourceScan.scala?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
move original class 'CarbonDataSourceScan' to src path 'commonTo2.1And2.2', and add a new class 'CarbonDataSourceScan' in src path 'spark2.3' which is added some lazy parameters.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My comment : only for 4 parameters , copy the whole file(CarbonDataSourceScan.scala) for spark 2.3 integration, may not require. see if can add the judgement for different spark version with different code/parameters.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@chenliang613 , it's not about the different code/parameters, it changes common val parameters to lazy val parameters. I think it's difficult to add lazy keyword on val parameters according to spark version.
Or does anyone know how to do and give me some help, thanks.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
ok.
|
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/619/ |
|
Build Failed with Spark 2.3.1, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/8880/ |
|
Build Success with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/811/ |
|
@zzcclp Please check and fix the tests |
|
@ravipesala can you help me to check why these three test cases fail? It's about the decimal precision. |
|
@ravipesala I know how to fix and will fix the tests ASAP. |
Upgrade spark integration version to 2.3.2
586cf7b to
c8b53cf
Compare
|
Build Success with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/837/ |
|
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/643/ |
|
Build Success with Spark 2.3.1, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/8905/ |
|
@sujith71955 @chenliang613 @ravipesala @jackylk this pr is ready, please review, thanks. |
|
My comment : only for 4 parameters , copy the whole file(CarbonDataSourceScan.scala) for spark 2.3 integration, may not require. see if can add the judgement for different spark version with different code/parameters. |
|
I think it is not possible to add lazy variable without breaking spark 2.2 integration. Since this class is small, I think it is ok to have a separate one for spark 2.3 integration |
|
@sujith71955 any suggestion for this? |
|
LGTM |
1 similar comment
|
LGTM |
|
LGTM Even though SPARK-PR#21815 changes is a workaround solution and in future its subjected to change. Need a caution here. Can you please add this PR reference in the modified code of the carbon for future reference. I think you add like ' made lazy since spark 2.3.2 version(SPARK-PR#21815) |
|
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/656/ |
|
Build Success with Spark 2.3.1, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/8920/ |
|
Build Failed with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/852/ |
|
the failed test case is not related to this pr, right? |
|
I think better to re-trigger once |
|
retest this please |
|
Seems to be impact of this PR , but it has passed previously,since we modified the outputpartition as lazy, better to have a relook once |
|
ok, but it doesn't add lazy for spak 2.2. |
|
Right, independently run this case and check once. Also retrigger the
build. Thanks
…On Sun, 30 Sep 2018 at 12:44 PM, Zhichao Zhang ***@***.***> wrote:
ok, but it doesn't add lazy for spak 2.2.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#2779 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AMZZ-TP2bfR9iJTUhBM1yDJDUusA5VcDks5ugG9HgaJpZM4W9IBA>
.
|
|
Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder2.1/660/ |
|
Build Success with Spark 2.3.1, Please check CI http://136.243.101.176:8080/job/carbondataprbuilder2.3/8923/ |
|
Build Success with Spark 2.2.1, Please check CI http://95.216.28.178:8080/job/ApacheCarbonPRBuilder1/855/ |
Changes:
Be sure to do all of the following checklist to help us incorporate
your contribution quickly and easily:
Any interfaces changed? No
Any backward compatibility impacted? No
Document update required? No
Testing done
Please provide details on
- Whether new unit test cases have been added or why no new tests are required?
- How it is tested? Please attach test report.
- Is it a performance related change? Please attach the performance test report.
- Any additional information to help reviewers in testing this change.
For large changes, please consider breaking it into sub-tasks under an umbrella JIRA.