summaryrefslogtreecommitdiffstats
path: root/tests/test_parser.py
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2023-06-02 23:59:40 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2023-06-02 23:59:46 +0000
commit20739a12c39121a9e7ad3c9a2469ec5a6876199d (patch)
treec000de91c59fd29b2d9beecf9f93b84e69727f37 /tests/test_parser.py
parentReleasing debian version 12.2.0-1. (diff)
downloadsqlglot-20739a12c39121a9e7ad3c9a2469ec5a6876199d.tar.xz
sqlglot-20739a12c39121a9e7ad3c9a2469ec5a6876199d.zip
Merging upstream version 15.0.0.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'tests/test_parser.py')
-rw-r--r--tests/test_parser.py83
1 files changed, 67 insertions, 16 deletions
diff --git a/tests/test_parser.py b/tests/test_parser.py
index 816471e..84ae0b5 100644
--- a/tests/test_parser.py
+++ b/tests/test_parser.py
@@ -15,14 +15,23 @@ class TestParser(unittest.TestCase):
self.assertIsInstance(parse_one("left join foo", into=exp.Join), exp.Join)
self.assertIsInstance(parse_one("int", into=exp.DataType), exp.DataType)
self.assertIsInstance(parse_one("array<int>", into=exp.DataType), exp.DataType)
+ self.assertIsInstance(parse_one("foo", into=exp.Table), exp.Table)
+
+ with self.assertRaises(ParseError) as ctx:
+ parse_one("SELECT * FROM tbl", into=exp.Table)
+
+ self.assertEqual(
+ str(ctx.exception),
+ "Failed to parse 'SELECT * FROM tbl' into <class 'sqlglot.expressions.Table'>",
+ )
def test_parse_into_error(self):
- expected_message = "Failed to parse into [<class 'sqlglot.expressions.From'>]"
+ expected_message = "Failed to parse 'SELECT 1;' into [<class 'sqlglot.expressions.From'>]"
expected_errors = [
{
"description": "Invalid expression / Unexpected token",
"line": 1,
- "col": 7,
+ "col": 6,
"start_context": "",
"highlight": "SELECT",
"end_context": " 1;",
@@ -30,17 +39,18 @@ class TestParser(unittest.TestCase):
}
]
with self.assertRaises(ParseError) as ctx:
- parse_one("SELECT 1;", "sqlite", [exp.From])
+ parse_one("SELECT 1;", read="sqlite", into=[exp.From])
+
self.assertEqual(str(ctx.exception), expected_message)
self.assertEqual(ctx.exception.errors, expected_errors)
def test_parse_into_errors(self):
- expected_message = "Failed to parse into [<class 'sqlglot.expressions.From'>, <class 'sqlglot.expressions.Join'>]"
+ expected_message = "Failed to parse 'SELECT 1;' into [<class 'sqlglot.expressions.From'>, <class 'sqlglot.expressions.Join'>]"
expected_errors = [
{
"description": "Invalid expression / Unexpected token",
"line": 1,
- "col": 7,
+ "col": 6,
"start_context": "",
"highlight": "SELECT",
"end_context": " 1;",
@@ -49,7 +59,7 @@ class TestParser(unittest.TestCase):
{
"description": "Invalid expression / Unexpected token",
"line": 1,
- "col": 7,
+ "col": 6,
"start_context": "",
"highlight": "SELECT",
"end_context": " 1;",
@@ -58,6 +68,7 @@ class TestParser(unittest.TestCase):
]
with self.assertRaises(ParseError) as ctx:
parse_one("SELECT 1;", "sqlite", [exp.From, exp.Join])
+
self.assertEqual(str(ctx.exception), expected_message)
self.assertEqual(ctx.exception.errors, expected_errors)
@@ -75,7 +86,7 @@ class TestParser(unittest.TestCase):
def test_table(self):
tables = [t.sql() for t in parse_one("select * from a, b.c, .d").find_all(exp.Table)]
- self.assertEqual(tables, ["a", "b.c", "d"])
+ self.assertEqual(set(tables), {"a", "b.c", "d"})
def test_union_order(self):
self.assertIsInstance(parse_one("SELECT * FROM (SELECT 1) UNION SELECT 2"), exp.Union)
@@ -92,7 +103,7 @@ class TestParser(unittest.TestCase):
self.assertEqual(len(parse_one("select * from (select 1) x cross join y").args["joins"]), 1)
self.assertEqual(
parse_one("""SELECT * FROM x CROSS JOIN y, z LATERAL VIEW EXPLODE(y)""").sql(),
- """SELECT * FROM x, z CROSS JOIN y LATERAL VIEW EXPLODE(y)""",
+ """SELECT * FROM x CROSS JOIN y, z LATERAL VIEW EXPLODE(y)""",
)
self.assertIsNone(
parse_one("create table a as (select b from c) index").find(exp.TableAlias)
@@ -156,8 +167,8 @@ class TestParser(unittest.TestCase):
assert expression.expressions[2].alias == "c"
assert expression.expressions[3].alias == "D"
assert expression.expressions[4].alias == "y|z'"
- table = expression.args["from"].expressions[0]
- assert table.this.name == "z"
+ table = expression.args["from"].this
+ assert table.name == "z"
assert table.args["db"].name == "y"
def test_multi(self):
@@ -168,8 +179,8 @@ class TestParser(unittest.TestCase):
)
assert len(expressions) == 2
- assert expressions[0].args["from"].expressions[0].this.name == "a"
- assert expressions[1].args["from"].expressions[0].this.name == "b"
+ assert expressions[0].args["from"].name == "a"
+ assert expressions[1].args["from"].name == "b"
expressions = parse("SELECT 1; ; SELECT 2")
@@ -202,6 +213,15 @@ class TestParser(unittest.TestCase):
with self.assertRaises(ParseError):
parse_one("WITH cte AS (SELECT * FROM x)")
+ self.assertEqual(
+ parse_one(
+ "CREATE TABLE t (i UInt8) ENGINE = AggregatingMergeTree() ORDER BY tuple()",
+ read="clickhouse",
+ error_level=ErrorLevel.RAISE,
+ ).sql(dialect="clickhouse"),
+ "CREATE TABLE t (i UInt8) ENGINE=AggregatingMergeTree() ORDER BY tuple()",
+ )
+
def test_space(self):
self.assertEqual(
parse_one("SELECT ROW() OVER(PARTITION BY x) FROM x GROUP BY y").sql(),
@@ -292,6 +312,7 @@ class TestParser(unittest.TestCase):
self.assertIsInstance(parse_one("TIMESTAMP('2022-01-01')"), exp.Func)
self.assertIsInstance(parse_one("TIMESTAMP()"), exp.Func)
self.assertIsInstance(parse_one("map.x"), exp.Column)
+ self.assertIsInstance(parse_one("CAST(x AS CHAR(5))").to.expressions[0], exp.DataTypeSize)
def test_set_expression(self):
set_ = parse_one("SET")
@@ -415,39 +436,56 @@ class TestParser(unittest.TestCase):
) PIVOT (AVG(price), MAX(quality) FOR partname IN ('prop' AS prop1, 'rudder'))
"""
- multiple_aggregates_not_aliased_with_quoted_identifier = """
+ multiple_aggregates_not_aliased_with_quoted_identifier_spark = """
SELECT * FROM (
SELECT partname, price, quality FROM part
) PIVOT (AVG(`PrIcE`), MAX(quality) FOR partname IN ('prop' AS prop1, 'rudder'))
"""
+ multiple_aggregates_not_aliased_with_quoted_identifier_duckdb = """
+ SELECT * FROM (
+ SELECT partname, price, quality FROM part
+ ) PIVOT (AVG("PrIcE"), MAX(quality) FOR partname IN ('prop' AS prop1, 'rudder'))
+ """
+
query_to_column_names = {
nothing_aliased: {
"bigquery": ["prop", "rudder"],
+ "duckdb": ["prop", "rudder"],
"redshift": ["prop", "rudder"],
- "snowflake": ['"prop"', '"rudder"'],
+ "snowflake": ['''"'prop'"''', '''"'rudder'"'''],
"spark": ["prop", "rudder"],
},
everything_aliased: {
"bigquery": ["avg_price_prop1", "avg_price_rudder1"],
+ "duckdb": ["prop1_avg_price", "rudder1_avg_price"],
"redshift": ["prop1_avg_price", "rudder1_avg_price"],
"spark": ["prop1", "rudder1"],
},
only_pivot_columns_aliased: {
"bigquery": ["prop1", "rudder1"],
+ "duckdb": ["prop1", "rudder1"],
"redshift": ["prop1", "rudder1"],
"spark": ["prop1", "rudder1"],
},
columns_partially_aliased: {
"bigquery": ["prop1", "rudder"],
+ "duckdb": ["prop1", "rudder"],
"redshift": ["prop1", "rudder"],
"spark": ["prop1", "rudder"],
},
multiple_aggregates_aliased: {
"bigquery": ["p_prop1", "q_prop1", "p_rudder", "q_rudder"],
+ "duckdb": ["prop1_p", "prop1_q", "rudder_p", "rudder_q"],
"spark": ["prop1_p", "prop1_q", "rudder_p", "rudder_q"],
},
multiple_aggregates_not_aliased: {
+ "duckdb": [
+ '"prop1_avg(price)"',
+ '"prop1_max(quality)"',
+ '"rudder_avg(price)"',
+ '"rudder_max(quality)"',
+ ],
"spark": [
"`prop1_avg(price)`",
"`prop1_max(quality)`",
@@ -455,7 +493,7 @@ class TestParser(unittest.TestCase):
"`rudder_max(quality)`",
],
},
- multiple_aggregates_not_aliased_with_quoted_identifier: {
+ multiple_aggregates_not_aliased_with_quoted_identifier_spark: {
"spark": [
"`prop1_avg(PrIcE)`",
"`prop1_max(quality)`",
@@ -463,10 +501,23 @@ class TestParser(unittest.TestCase):
"`rudder_max(quality)`",
],
},
+ multiple_aggregates_not_aliased_with_quoted_identifier_duckdb: {
+ "duckdb": [
+ '"prop1_avg(PrIcE)"',
+ '"prop1_max(quality)"',
+ '"rudder_avg(PrIcE)"',
+ '"rudder_max(quality)"',
+ ],
+ },
}
for query, dialect_columns in query_to_column_names.items():
for dialect, expected_columns in dialect_columns.items():
expr = parse_one(query, read=dialect)
- columns = expr.args["from"].expressions[0].args["pivots"][0].args["columns"]
+ columns = expr.args["from"].this.args["pivots"][0].args["columns"]
self.assertEqual(expected_columns, [col.sql(dialect=dialect) for col in columns])
+
+ def test_parse_properties(self):
+ self.assertEqual(
+ parse_one("create materialized table x").sql(), "CREATE MATERIALIZED TABLE x"
+ )