Skip to content

Commit

Permalink
fix default order by
Browse files Browse the repository at this point in the history
  • Loading branch information
sh-rp committed Apr 2, 2024
1 parent b0434a6 commit 8f0f2b6
Show file tree
Hide file tree
Showing 3 changed files with 9 additions and 15 deletions.
8 changes: 6 additions & 2 deletions dlt/destinations/sql_jobs.py
Original file line number Diff line number Diff line change
Expand Up @@ -194,7 +194,7 @@ def gen_key_table_clauses(

@classmethod
def gen_delete_temp_table_sql(
cls, unique_column: str, key_table_clauses: Sequence[str], sql_client: SqlClientBase[Any]
cls, unique_column: str, key_table_clauses: Sequence[str], sql_client: SqlClientBase[Any]
) -> Tuple[List[str], str]:
"""Generate sql that creates delete temp table and inserts `unique_column` from root table for all records to delete. May return several statements.
Expand Down Expand Up @@ -254,7 +254,7 @@ def gen_select_from_dedup_sql(
1) To select the values for an INSERT INTO statement.
2) To select the values for a temporary table used for inserts.
"""
order_by = "NULL"
order_by = cls.default_order_by()
if dedup_sort is not None:
order_by = f"{dedup_sort[0]} {dedup_sort[1].upper()}"
if condition is None:
Expand All @@ -271,6 +271,10 @@ def gen_select_from_dedup_sql(
) AS _dlt_dedup_numbered WHERE _dlt_dedup_rn = 1 AND ({condition})
"""

@classmethod
def default_order_by(cls) -> str:
return "(SELECT NULL)"

@classmethod
def gen_insert_temp_table_sql(
cls,
Expand Down
5 changes: 1 addition & 4 deletions tests/load/pipeline/test_pipelines.py
Original file line number Diff line number Diff line change
Expand Up @@ -58,10 +58,7 @@ def test_default_pipeline_names(
p.config.use_single_dataset = use_single_dataset
# this is a name of executing test harness or blank pipeline on windows
possible_names = ["dlt_pytest", "dlt_pipeline"]
possible_dataset_names = [
"dlt_pytest_dataset",
"dlt_pipeline_dataset"
]
possible_dataset_names = ["dlt_pytest_dataset", "dlt_pipeline_dataset"]
assert p.pipeline_name in possible_names
assert p.pipelines_dir == os.path.abspath(os.path.join(TEST_STORAGE_ROOT, ".dlt", "pipelines"))
assert p.dataset_name in possible_dataset_names
Expand Down
11 changes: 2 additions & 9 deletions tests/load/pipeline/test_write_disposition_changes.py
Original file line number Diff line number Diff line change
Expand Up @@ -133,10 +133,7 @@ def source():
return

# without a root key this will fail, it is expected
if (
not with_root_key
and destination_config.supports_merge
):
if not with_root_key and destination_config.supports_merge:
with pytest.raises(PipelineStepFailed):
pipeline.run(
s,
Expand All @@ -157,11 +154,7 @@ def source():
pipeline,
{
"items": 100 if destination_config.supports_merge else 200,
"items__sub_items": (
100
if destination_config.supports_merge
else 200
),
"items__sub_items": 100 if destination_config.supports_merge else 200,
},
)
assert pipeline.default_schema._normalizers_config["json"]["config"]["propagation"]["tables"][
Expand Down

0 comments on commit 8f0f2b6

Please sign in to comment.