-
Notifications
You must be signed in to change notification settings - Fork 29.1k
[SPARK-16525][SQL] Enable Row Based HashMap in HashAggregateExec #14176
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 1 commit
4a5b81f
7194394
122cf18
def94cc
97bb7c1
b9a4268
e67ff5d
b32cb7b
a58314c
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
…orized hashmap
- Loading branch information
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -499,14 +499,16 @@ object SQLConf { | |
| .intConf | ||
| .createWithDefault(40) | ||
|
|
||
| val VECTORIZED_AGG_MAP_MAX_COLUMNS = | ||
| SQLConfigBuilder("spark.sql.codegen.aggregate.map.columns.max") | ||
| val ENFORCE_FAST_AGG_MAP_IMPL = | ||
| SQLConfigBuilder("spark.sql.codegen.aggregate.map.enforce.impl") | ||
|
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Let's also make sure that all references to the old config are also appropriately modified. |
||
| .internal() | ||
| .doc("Sets the maximum width of schema (aggregate keys + values) for which aggregate with" + | ||
| "keys uses an in-memory columnar map to speed up execution. Setting this to 0 effectively" + | ||
| "disables the columnar map") | ||
| .intConf | ||
| .createWithDefault(3) | ||
| .doc("Sets the implementation for fast hash map during aggregation. Could be one of the " + | ||
| "following: rowbased, vectorized, skip, auto. Defaults to auto, and should only be other " + | ||
| "values for testing purposes.") | ||
| .stringConf | ||
| .transform(_.toLowerCase()) | ||
| .checkValues(Set("rowbased", "vectorized", "skip", "auto")) | ||
| .createWithDefault("auto") | ||
|
||
|
|
||
| val FILE_SINK_LOG_DELETION = SQLConfigBuilder("spark.sql.streaming.fileSink.log.deletion") | ||
| .internal() | ||
|
|
@@ -673,7 +675,7 @@ private[sql] class SQLConf extends Serializable with CatalystConf with Logging { | |
|
|
||
| override def runSQLonFile: Boolean = getConf(RUN_SQL_ON_FILES) | ||
|
|
||
| def vectorizedAggregateMapMaxColumns: Int = getConf(VECTORIZED_AGG_MAP_MAX_COLUMNS) | ||
| def enforceFastAggHashMapImpl: String = getConf(ENFORCE_FAST_AGG_MAP_IMPL) | ||
|
|
||
| def variableSubstituteEnabled: Boolean = getConf(VARIABLE_SUBSTITUTE_ENABLED) | ||
|
|
||
|
|
||
Uh oh!
There was an error while loading. Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: Maybe just call this
FAST_AGG_MAP_IMPL?