You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is a heads-up as to what is coming in the future.
At GA the DCP will likely demonstrate proof-of-concept for image-based transcriptomics using the SpaceTx-Allen dataset, which is 600 GB in size.
Valdiation
If there is a validator written for image files (need to confirm), upload currently has a size limit of 1TB for file validation. The aggregate size of all files listed in 1 validation request cannot exceed 1TB.
Aside: we need to look at the number of validation requests assigned to each batch worker. We could easily blow the 1TB limit if several validation requests are assigned to the same server.
If there is a validator written for images, Upload, as currently designed, can support the SpaceTx-Allen dataset.
However, the imaging team also tells us: "Imaging datasets are estimated to be up to 50 TB in size, based on literature and existing example datasets.". Upload cannot validate images of that size, at this time. Of the top of my head, there are several tactics we could use to overcome this limitation:
Increase size of volume attached to validator AMI image to 50TB - unappealing.
@sampierson As it stands, no work is required for current approach to ingest/upload spacetx data. Closing this ticket and new ticket will be created if this changes
This is a heads-up as to what is coming in the future.
At GA the DCP will likely demonstrate proof-of-concept for image-based transcriptomics using the SpaceTx-Allen dataset, which is 600 GB in size.
Valdiation
If there is a validator written for image files (need to confirm), upload currently has a size limit of 1TB for file validation. The aggregate size of all files listed in 1 validation request cannot exceed 1TB.
Aside: we need to look at the number of validation requests assigned to each batch worker. We could easily blow the 1TB limit if several validation requests are assigned to the same server.
If there is a validator written for images, Upload, as currently designed, can support the SpaceTx-Allen dataset.
However, the imaging team also tells us: "Imaging datasets are estimated to be up to 50 TB in size, based on literature and existing example datasets.". Upload cannot validate images of that size, at this time. Of the top of my head, there are several tactics we could use to overcome this limitation:
The text was updated successfully, but these errors were encountered: