-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Missing Observations #176
Comments
I checked it out, and the perimeters I get from the API are the same as those I get from both the combined_largefires and the allfires files read directly from s3. I had been thinking that maybe something was wrong with how we save off the largefires file from allfires during postprocessing, but this does not support that idea. It also doesn't support the idea that data is being mangled somewhere in the API upload pipeline. |
Yes, I think you've found it. We never fully fixed this issue --- that we essentially need to run the algorithm multiple times a day regenerating files until maybe 12 hours later to account for time differences and upload differences across the US. |
This is especially annoying because it seems like it's sensor dependent --- and we aren't certain how variable the upload timing is. |
@zebbecker @mccabete, thanks for diving into this. Really a frustrating issue. Let me see if I'm interpreting this correctly:
I'm trying to remember how we're currently doing the NRT runs, but could we swap the run instructions to make |
@eorland huh that's an interesting thought! I think it's a good idea, with the one detail that right now I think the algorithm is very aggressive about not regenerating files it doesn't need to. I think we'd need to tweak that so that if the file is close in time to the present (-24 hours like you suggest) the algorithm is ok with regenerating it -- I'm not sure if this will clog up our runs though. |
Copying Yang's input here to have it all in one place for future reference: "So based on discussions in the github issue page, it seems those tracking errors were mainly due to the operational FEDS run before the complete recording of VIIRS active fires in the FIRMS files. For most of the cases, the current time lag setting (can you remind me the current UTC/local time scheduled to run the FEDS code?) ensures the availability of VIIRS data. But sometimes the FIRMS data may be delayed and FEDS code may use incorrect/incomplete data. While we are unable to correct this for the latest time step, we may fix it by performing a 'remedy' run for the previous time step (assuming the data for t-12hrs have been complete at the current time). To make it more efficient, we may also do an initial check to see if the fire pixels for t-12hr from FIRMS are different from the preprocessed local data. If there's no change, the 'remedy' run may not be necessary. If there's change, we can extract the VIIRS data for t-12hr, re-run the code, and overwrite the output (from the previous 'normal' run) for that time step. The backward time span for this kind of 'remedy' run may be extended to more than 1 time step, considering the possibility of longer FIRMS data delay. While this is indeed an upstream data issue, we probably can minimize the negative effect in this way. BTW, does the current operational run only use NOAA 20 data?" |
To answer specific questions:
In general: I like the idea of essentially re-running (forcibly) firetracking from time t-N up to t when we process t. This would help provide additional resiliency to unexpected delays in data downlinking that are outside of our control. If we assume that the full record might not be available when we run for the first time, and therefore are prepared to update our fire tracking with subsequent observations, we can also move our initial NRT runs closer to actual overpass times, reducing latency by several hours in the best case. As long as we incorporate Yang's suggestion of checking if a "remedy run" is actually necessary before doing them, I suspect that we will not significantly decrease performance/clog up runs with this strategy, as we won't need to recompute most timesteps at all. Only when something has gone wrong. |
@zebbecker and @eorland noticed that some of the fires this year seemed to be growing only every 24 hours. Zeb made the below figure.
The text was updated successfully, but these errors were encountered: