Skip to content
Closed
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Fix verbatims
  • Loading branch information
hvanhovell committed Jul 25, 2023
commit 751fc872dd0fb2fe008e4efe8f86e96aa8891550
Original file line number Diff line number Diff line change
Expand Up @@ -239,195 +239,84 @@ DescribeTableCommand `spark_catalog`.`default`.`char_tbl1`, true, [col_name#x, d
-- !query
create table char_part(c1 char(5), c2 char(2), v1 varchar(6), v2 varchar(2)) using parquet partitioned by (v2, c2)
-- !query analysis
org.apache.spark.SparkRuntimeException
org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException
{
"errorClass" : "LOCATION_ALREADY_EXISTS",
"sqlState" : "42710",
"errorClass" : "TABLE_OR_VIEW_ALREADY_EXISTS",
"sqlState" : "42P07",
"messageParameters" : {
"identifier" : "`spark_catalog`.`default`.`char_part`",
"location" : "'file:/Users/herman/Engineering/spark-dev/sql/core/spark-warehouse/org.apache.spark.sql.SQLQueryTestSuite/char_part'"
"relationName" : "`spark_catalog`.`default`.`char_part`"
}
}


-- !query
desc formatted char_part
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 16,
"stopIndex" : 24,
"fragment" : "char_part"
} ]
}
DescribeTableCommand `spark_catalog`.`default`.`char_part`, true, [col_name#x, data_type#x, comment#x]


-- !query
alter table char_part add partition (v2='ke', c2='nt') location 'loc1'
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
org.apache.spark.sql.catalyst.analysis.PartitionsAlreadyExistException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"errorClass" : "PARTITIONS_ALREADY_EXIST",
"sqlState" : "428FT",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 13,
"stopIndex" : 21,
"fragment" : "char_part"
} ]
"partitionList" : "PARTITION (`v2` = ke, `c2` = nt)",
"tableName" : "`default`.`char_part`"
}
}


-- !query
desc formatted char_part
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 16,
"stopIndex" : 24,
"fragment" : "char_part"
} ]
}
DescribeTableCommand `spark_catalog`.`default`.`char_part`, true, [col_name#x, data_type#x, comment#x]


-- !query
alter table char_part partition (v2='ke') rename to partition (v2='nt')
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
org.apache.spark.sql.AnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"errorClass" : "_LEGACY_ERROR_TEMP_1232",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 13,
"stopIndex" : 21,
"fragment" : "char_part"
} ]
"partitionColumnNames" : "v2, c2",
"specKeys" : "v2",
"tableName" : "`spark_catalog`.`default`.`char_part`"
}
}


-- !query
desc formatted char_part
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 16,
"stopIndex" : 24,
"fragment" : "char_part"
} ]
}
DescribeTableCommand `spark_catalog`.`default`.`char_part`, true, [col_name#x, data_type#x, comment#x]


-- !query
alter table char_part partition (v2='ke', c2='nt') set location 'loc2'
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 13,
"stopIndex" : 21,
"fragment" : "char_part"
} ]
}
AlterTableSetLocationCommand `spark_catalog`.`default`.`char_part`, Map(v2 -> ke, c2 -> nt), loc2


-- !query
desc formatted char_part
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 16,
"stopIndex" : 24,
"fragment" : "char_part"
} ]
}
DescribeTableCommand `spark_catalog`.`default`.`char_part`, true, [col_name#x, data_type#x, comment#x]


-- !query
MSCK REPAIR TABLE char_part
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 19,
"stopIndex" : 27,
"fragment" : "char_part"
} ]
}
RepairTableCommand `spark_catalog`.`default`.`char_part`, true, false, MSCK REPAIR TABLE


-- !query
desc formatted char_part
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`char_part`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 16,
"stopIndex" : 24,
"fragment" : "char_part"
} ]
}
DescribeTableCommand `spark_catalog`.`default`.`char_part`, true, [col_name#x, data_type#x, comment#x]


-- !query
Expand Down
30 changes: 6 additions & 24 deletions sql/core/src/test/resources/sql-tests/analyzer-results/hll.sql.out
Original file line number Diff line number Diff line change
Expand Up @@ -9,35 +9,17 @@ DropTable true, false
-- !query
CREATE TABLE t1 USING JSON AS VALUES (0), (1), (2), (2), (2), (3), (4) as tab(col)
-- !query analysis
org.apache.spark.SparkRuntimeException
{
"errorClass" : "LOCATION_ALREADY_EXISTS",
"sqlState" : "42710",
"messageParameters" : {
"identifier" : "`spark_catalog`.`default`.`t1`",
"location" : "'file:/Users/herman/Engineering/spark-dev/sql/core/spark-warehouse/org.apache.spark.sql.SQLQueryTestSuite/t1'"
}
}
CreateDataSourceTableAsSelectCommand `spark_catalog`.`default`.`t1`, ErrorIfExists, [col]
+- SubqueryAlias tab
+- LocalRelation [col#x]


-- !query
SELECT hll_sketch_estimate(hll_sketch_agg(col)) AS result FROM t1
-- !query analysis
org.apache.spark.sql.catalyst.ExtendedAnalysisException
{
"errorClass" : "TABLE_OR_VIEW_NOT_FOUND",
"sqlState" : "42P01",
"messageParameters" : {
"relationName" : "`t1`"
},
"queryContext" : [ {
"objectType" : "",
"objectName" : "",
"startIndex" : 64,
"stopIndex" : 65,
"fragment" : "t1"
} ]
}
Aggregate [hll_sketch_estimate(hll_sketch_agg(col#x, 12, 0, 0)) AS result#xL]
+- SubqueryAlias spark_catalog.default.t1
+- Relation spark_catalog.default.t1[col#x] json


-- !query
Expand Down
Loading