-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[native] int64 timestamp column in Parquet overflowing to large values #23769
Comments
@ethanyzhang imported this issue into IBM GitHub Enterprise |
@kishansairam9 What is the data type for start_time_millis_ts? timestamp(3)? Do you have the sample data that we can reproduce it? |
datatype is of timestamp with no time zone |
@yingsu00 I shared the Parquet files with Zuyu. |
@kishansairam9 the root cause is that velox Parquet reader lacks of supports for INT64 timestamp. Internally a timestamp value treated as |
Fixed by facebookincubator/velox#11530 |
Running the following query failed multiple times with too large timestamp error
The exact same query on same datasource finished on Trino without any issues, so no corruption in data, something is happening in presto native layer.
Your Environment
Expected Behavior
Query finishes with no error and returns columns
Current Behavior
Fails with timestamp conversion as too large since epoch. Each time a different seconds from epoch is provided with large values. Probably initialisation or invalid access issue.
Steps to Reproduce
I have shared a sample data from the iceberg partition which this query should act on to @majetideepak, cannot attach it here on public forum. Please reach out over Presto Slack
@Kishan
if you would like to investigate this.Context
We are experimenting with presto native to improve our SLAs on datalake. This failure of simple query doesn't inspire confidence for switching in near future.
The text was updated successfully, but these errors were encountered: