-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Backup not functional #2270
Comments
Following the docs, fixing the backup instance fails (after 22 minutes 😱)
Possibly because my user is missing some backup roles, I am an Owner and Storage Blob Data Owner though. |
The problem may be that the backup vault doesn't have the correct role assigned in the target storage account. |
Assigning the needed role to the backup vault fixed that issue. Next problem is
|
Storage account kind is BlockBlobStorage
from here |
Is this the correct storage account. The target is sensitivedata which has the inputs and outputs. That feels like the least important thing to backup. In fact, I could see a strong case for not making copies of the input data when you are acting as a data processor. |
I don't think we're ever acting as a data processor - pretty much all useful research is going to involve the researchers making significant-enough decisions that they are a data controller. @Davsarper : can you remember what we should be backing up for DSPT-compatibility? |
Agreed on us being data controllers. So we refer to backups when asked about our business continuity plan, and currently answer that after failure we would recover 'as much as possible'. I don't see on first instance a hard description of what needs to be recovered for our organisation. I think critical things to recover would be those necessary for offering whichever (healthcare) services, since we do not then it might be relatively up to us to decide what is key to backup. I will have a read on business continuity plans, which we should develop https://www.digitalcarehub.co.uk/resource/creating-and-testing-a-business-continuity-plan-for-data-and-cyber-security/ |
I think we should be careful not to focus too much on us. I think the use case for a ephemeral TRE strictly for data processing is strong. I feel we might have tried to backup the inverse of what we really want,
|
Also,
🫠 |
So, backing up either the block blobs or NFS shares will require either,
My feeling is backing up to some redundant storage using borg is the most flexible and easiest to implement. |
Maybe the best way to implement our own is,
That way you don't need to worry about multiple workspaces running conflicting backup jobs. |
@jemrobinson @craddm I think we need to reach a consensus for the change we want to make here. |
I would like to see all the data/state necessary to recreate the environment backed up. Imagine, for example, that an SRE has been compromised by a ransomware attack 4 years into a 5 year project. Would our backups be sufficient to deploy an identical (or updated) new environment that results in only days/weeks of work being lost? We should also ensure that we have tiered backups (e.g. X days of daily backups + Y weeks of weekly backups + Z months of monthly backups) in case of longer-term problems. I'm agnostic as to the method used to achieve this: using Azure built-in solutions would be nice, but it's more important to have something that works than something with a point-and-click interface. @martintoreilly do you agree with this? |
Agreed, in terms of borg+borgmatic as a solution, Backups are incremental, encrypted and hashed. That gives good protection over that data becoming corrupted (you can schedule regular checks) while reducing the space needed. For restoration/disaster recovery. I think what we need to backup is |
@craddm @jemrobinson Any objection before I start along the lines I've posted here? |
Sounds good to me. It would be worth thinking how we could backup/restore |
For container instances, it is possible to mount Azure file shares but not blobs. |
This doesn't work for NFS file shares - it's just the standard kind that you can browse in the portal or with Storage Explorer. |
In that case, I think the best way forward is to add a VM and configure with cloud-init and Ansible. The workload should fit a small burstable size. |
✅ Checklist
💻 System information
📦 Packages
List of packages
🚫 Describe the problem
Backup not functional, workaround in docs doesn't work.
A major problem here is, the storage account types we have chosen (for good reasons of performance, ability to upload/download to/from) are not compatible with Azure backup services.
🚂 Workarounds or solutions
Because of the incompatibility, I think our two options are
The text was updated successfully, but these errors were encountered: