We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
In spark delta table you can enable an option to manage out of range versions or timestamps. https://docs.delta.io/latest/delta-change-data-feed.html#read-changes-in-streaming-queries
Right now the behaviour of load_cdf is inconsistent, if you provide an out of range version you get an error:
But with a timestamp out of range, you get an empty dataset:
It would be useful for incremental pipelines to have a way to manage this behaviour and make it consistent.
The text was updated successfully, but these errors were encountered:
If you know some rust it's probably a simple fix
Sorry, something went wrong.
@ion-elgreco I have never used rust and it is been a while since I have done anything that needs to be compiles, but can give it a go
Successfully merging a pull request may close this issue.
Description
In spark delta table you can enable an option to manage out of range versions or timestamps. https://docs.delta.io/latest/delta-change-data-feed.html#read-changes-in-streaming-queries
Right now the behaviour of load_cdf is inconsistent, if you provide an out of range version you get an error:
But with a timestamp out of range, you get an empty dataset:
It would be useful for incremental pipelines to have a way to manage this behaviour and make it consistent.
The text was updated successfully, but these errors were encountered: