Replies: 1 comment 1 reply
-
That's correct.
That's correct.
As stated in Algorithm 1 in the paper, TSMixup augmentations were generated with length uniformly sampled between 128 and 2048. That gives an average length of 1088. Therefore approximately 1088 * 10^7 observations were generated with this process. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi @abdulfatir ,
I would like to ask how many data chronos used for pretraining and how to get the numbers stated in the paper. As mentioned in Sec 5.1, 890K univariate time series are used (in total roughly 84B tokens). Are these numbers for original raw data but not directly for training?
In Sec 5.2, it states 10M TSMixup augmentations and 1M synthetic time series (so total 11M time series). In my view, these numbers are the data used for training. Am I correct?
Besides, I would like to ask how many observations/tokens are used for training given that we are using 11M time series? I noticed that synthetic time series have 1024 tokens each as shown in https://github.com/amazon-science/chronos-forecasting/blob/main/scripts/kernel-synth.py. But how about TSMixup augmentations?
Beta Was this translation helpful? Give feedback.
All reactions