If you are using a released version of Kubernetes, you should refer to the docs that go with that version.
The latest release of this document can be found [here](http://releases.k8s.io/release-1.4/docs/design/volume-snapshotting.md).Documentation for other releases can be found at releases.k8s.io.
Authors: Cindy Wang
Many storage systems (GCE PD, Amazon EBS, etc.) provide the ability to create "snapshots" of a persistent volumes to protect against data loss. Snapshots can be used in place of a traditional backup system to back up and restore primary and critical data. Snapshots allow for quick data backup (for example, it takes a fraction of a second to create a GCE PD snapshot) and offer fast recovery time objectives (RTOs) and recovery point objectives (RPOs).
Typical existing backup solutions offer on demand or scheduled snapshots.
An application developer using a storage may want to create a snapshot before an update or other major event. Kubernetes does not currently offer a standardized snapshot API for creating, listing, deleting, and restoring snapshots on an arbitrary volume.
Existing solutions for scheduled snapshotting include cron jobs and external storage drivers. Some cloud storage volumes can be configured to take automatic snapshots, but this is specified on the volumes themselves.
For the first version of snapshotting support in Kubernetes, only on-demand snapshots will be supported. Features listed in the roadmap for future versions are also nongoals.
-
Goal 1: Enable on-demand snapshots of Kubernetes persistent volumes by application developers.
- Nongoal: Enable automatic periodic snapshotting for direct volumes in pods.
-
Goal 2: Expose standardized snapshotting operations Create and List in Kubernetes REST API.
- Nongoal: Support Delete and Restore snapshot operations in API.
-
Goal 3: Implement snapshotting interface for GCE PDs.
- Nongoal: Implement snapshotting interface for non GCE PD volumes.
Major features, in order of priority (bold features are priorities for v1):
-
On demand snapshots
-
API to create new snapshots and list existing snapshots
-
API to restore a disk from a snapshot and delete old snapshots
-
-
Scheduled snapshots
-
Support snapshots for non-cloud storage volumes (i.e. plugins that require actions to be triggered from the node)
-
Time SLA from issuing a snapshot to completion:
-
The period we are interested is the time between the scheduled snapshot time and the time the snapshot is finishes uploading to its storage location.
-
This should be on the order of a few minutes.
-
Data corruption
-
Though it is generally recommended to stop application writes before executing the snapshot command, we will not do this for several reasons:
-
GCE and Amazon can create snapshots while the application is running.
-
Stopping application writes cannot be done from the master and varies by application, so doing so will introduce unnecessary complexity and permission issues in the code.
-
Most file systems and server applications are (and should be) able to restore inconsistent snapshots the same way as a disk that underwent an unclean shutdown.
-
-
-
Snapshot failure
-
Case: Failure during external process, such as during API call or upload
- Log error, retry until success (indefinitely)
-
Case: Failure within Kubernetes, such as controller restarts
- If the master restarts in the middle of a snapshot operation, then the controller does not know whether or not the operation succeeded. However, since the annotation has not been deleted, the controller will retry, which may result in a crash loop if the first operation has not yet completed. This issue will not be addressed in the alpha version, but future versions will need to address it by persisting state.
-
Snapshot operations will be triggered by annotations on PVC API objects.
-
Create:
-
Key: create.snapshot.volume.alpha.kubernetes.io
-
Value: [snapshot name]
-
-
List:
-
Key: snapshot.volume.alpha.kubernetes.io/[snapshot name]
-
Value: [snapshot timestamp]
-
A new controller responsible solely for snapshot operations will be added to the controllermanager on the master. This controller will watch the API server for new annotations on PVCs. When a create snapshot annotation is added, it will trigger the appropriate snapshot creation logic for the underlying persistent volume type. The list annotation will be populated by the controller and only identify all snapshots created for that PVC by Kubernetes.
The snapshot operation is a no-op for volume plugins that do not support snapshots via an API call (i.e. non-cloud storage).
-
Create snapshot
-
Usage:
-
Users create annotation with key "create.snapshot.volume.alpha.kubernetes.io", value does not matter
-
When the annotation is deleted, the operation has succeeded. The snapshot will be listed in the value of snapshot-list.
-
API is declarative and guarantees only that it will begin attempting to create the snapshot once the annotation is created and will complete eventually.
-
-
PVC control loop in master
-
If annotation on new PVC, search for PV of volume type that implements SnapshottableVolumePlugin. If one is available, use it. Otherwise, reject the claim and post an event to the PV.
-
If annotation on existing PVC, if PV type implements SnapshottableVolumePlugin, continue to SnapshotController logic. Otherwise, delete the annotation and post an event to the PV.
-
-
-
List existing snapshots
-
Only displayed as annotations on PVC object.
-
Only lists unique names and timestamps of snapshots taken using the Kubernetes API.
-
Usage:
-
Get the PVC object
-
Snapshots are listed as key-value pairs within the PVC annotations
-
-
PVC Informer: A shared informer that stores (references to) PVC objects, populated by the API server. The annotations on the PVC objects are used to add items to SnapshotRequests.
SnapshotRequests: An in-memory cache of incomplete snapshot requests that is populated by the PVC informer. This maps unique volume IDs to PVC objects. Volumes are added when the create snapshot annotation is added, and deleted when snapshot requests are completed successfully.
Reconciler: Simple loop that triggers asynchronous snapshots via the OperationExecutor. Deletes create snapshot annotation if successful.
The controller will have a loop that does the following:
-
Fetch State
- Fetch all PVC objects from the API server.
-
Act
-
Trigger snapshot:
- Loop through SnapshotRequests and trigger create snapshot logic (see below) for any PVCs that have the create snapshot annotation.
-
-
Persist State
- Once a snapshot operation completes, write the snapshot ID/timestamp to the PVC Annotations and delete the create snapshot annotation in the PVC object via the API server.
Snapshot operations can take a long time to complete, so the primary controller loop should not block on these operations. Instead the reconciler should spawn separate threads for these operations via the operation executor.
The controller will reject snapshot requests if the unique volume ID already exists in the SnapshotRequests. Concurrent operations on the same volume will be prevented by the operation executor.
To create a snapshot:
-
Acquire operation lock for volume so that no other attach or detach operations can be started for volume.
- Abort if there is already a pending operation for the specified volume (main loop will retry, if needed).
-
Spawn a new thread:
-
Execute the volume-specific logic to create a snapshot of the persistent volume reference by the PVC.
-
For any errors, log the error, and terminate the thread (the main controller will retry as needed).
-
Once a snapshot is created successfully:
-
Make a call to the API server to delete the create snapshot annotation in the PVC object.
-
Make a call to the API server to add the new snapshot ID/timestamp to the PVC Annotations.
-
-
Brainstorming notes below, read at your own risk!
Open questions:
-
What has more value: scheduled snapshotting or exposing snapshotting/backups as a standardized API?
-
It seems that the API route is a bit more feasible in implementation and can also be fully utilized.
- Can the API call methods on VolumePlugins? Yeah via controller
-
The scheduler gives users functionality that doesn’t already exist, but required adding an entirely new controller
-
-
Should the list and restore operations be part of v1?
-
Do we call them snapshots or backups?
- From the SIG email: "The snapshot should not be suggested to be a backup in any documentation, because in practice is is necessary, but not sufficient, when conducting a backup of a stateful application."
-
At what minimum granularity should snapshots be allowed?
-
How do we store information about the most recent snapshot in case the controller restarts?
-
In case of error, do we err on the side of fewer or more snapshots?
Snapshot Scheduler
- PVC API Object
A new field, backupSchedule, will be added to the PVC API Object. The value of this field must be a cron expression.
-
CRUD operations on snapshot schedules
-
Create: Specify a snapshot within a PVC spec as a cron expression
- The cron expression provides flexibility to decrease the interval between snapshots in future versions
-
Read: Display snapshot schedule to user via kubectl get pvc
-
Update: Do not support changing the snapshot schedule for an existing PVC
-
Delete: Do not support deleting the snapshot schedule for an existing PVC
- In v1, the snapshot schedule is tied to the lifecycle of the PVC. Update and delete operations are therefore not supported. In future versions, this may be done using kubectl edit pvc/name
-
-
Validation
-
Cron expressions must have a 0 in the minutes place and use exact, not interval syntax
- EBS appears to be able to take snapshots at the granularity of minutes, GCE PD takes at most minutes. Therefore for v1, we ensure that snapshots are taken at most hourly and at exact times (rather than at time intervals).
-
If Kubernetes cannot find a PV that supports snapshotting via its API, reject the PVC and display an error message to the user
-
Objective
Goal: Enable automatic periodic snapshotting (NOTE: A snapshot is a read-only copy of a disk.) for all kubernetes volume plugins.
Goal: Implement snapshotting interface for GCE PDs.
Goal: Protect against data loss by allowing users to restore snapshots of their disks.
Nongoal: Implement snapshotting support on Kubernetes for non GCE PD volumes.
Nongoal: Use snapshotting to provide additional features such as migration.
Background
Many storage systems (GCE PD, Amazon EBS, NFS, etc.) provide the ability to create "snapshots" of a persistent volumes to protect against data loss. Snapshots can be used in place of a traditional backup system to back up and restore primary and critical data. Snapshots allow for quick data backup (for example, it takes a fraction of a second to create a GCE PD snapshot) and offer fast recovery time objectives (RTOs) and recovery point objectives (RPOs).
Currently, no container orchestration software (i.e. Kubernetes and its competitors) provide snapshot scheduling for application storage.
Existing solutions for automatic snapshotting include cron jobs/shell scripts. Some volumes can be configured to take automatic snapshots, but this is specified on the volumes themselves, not via their associated applications. Snapshotting support gives Kubernetes clear competitive advantage for users who want automatic snapshotting on their volumes, and particularly those who want to configure application-specific schedules.
what is the value case? Who wants this? What do we enable by implementing this?
I think it introduces a lot of complexity, so what is the pay off? That should be clear in the document. Do mesos, or swarm or our competition implement this? AWS? Just curious.
Requirements
Functionality
Should this support PVs, direct volumes, or both?
Should we support deletion?
Should we support restores?
Automated schedule -- times or intervals? Before major event?
Performance
Snapshots are supposed to provide timely state freezing. What is the SLA from issuing one to it completing?
-
GCE: The snapshot operation takes a fraction of a second. If file writes can be paused, they should be paused until the snapshot is created (but can be restarted while it is pending). If file writes cannot be paused, the volume should be unmounted before snapshotting then remounted afterwards.
- Pending = uploading to GCE
-
EBS is the same, but if the volume is the root device the instance should be stopped before snapshotting
Reliability
How do we ascertain that deletions happen when we want them to?
For the same reasons that Kubernetes should not expose a direct create-snapshot command, it should also not allow users to delete snapshots for arbitrary volumes from Kubernetes.
We may, however, want to allow users to set a snapshotExpiryPeriod and delete snapshots once they have reached certain age. At this point we do not see an immediate need to implement automatic deletion (re:Saad) but may want to revisit this.
What happens when the snapshot fails as these are async operations?
Retry (for some time period? indefinitely?) and log the error
Other
What is the UI for seeing the list of snapshots?
In the case of GCE PD, the snapshots are uploaded to cloud storage. They are visible and manageable from the GCE console. The same applies for other cloud storage providers (i.e. Amazon). Otherwise, users may need to ssh into the device and access a ./snapshot or similar directory. In other words, users will continue to access snapshots in the same way as they have been while creating manual snapshots.
Overview
There are several design options for the design of each layer of implementation as follows.
- Public API:
Users will specify a snapshotting schedule for particular volumes, which Kubernetes will then execute automatically. There are several options for where this specification can happen. In order from most to least invasive:
1. New Volume API object
1. Currently, pods, PVs, and PVCs are API objects, but Volume is not. A volume is represented as a field within pod/PV objects and its details are lost upon destruction of its enclosing object.
2. We define Volume to be a brand new API object, with a snapshot schedule attribute that specifies the time at which Kubernetes should call out to the volume plugin to create a snapshot.
3. The Volume API object will be referenced by the pod/PV API objects. The new Volume object exists entirely independently of the Pod object.
4. Pros
1. Snapshot schedule conflicts: Since a single Volume API object ideally refers to a single volume, each volume has a single unique snapshot schedule. In the case where the same underlying PD is used by different pods which specify different snapshot schedules, we have a straightforward way of identifying and resolving the conflicts. Instead of using extra space to create duplicate snapshots, we can decide to, for example, use the most frequent snapshot schedule.
5. Cons
2. Heavyweight codewise; involves changing and touching a lot of existing code.
3. Potentially bad UX: How is the Volume API object created?
1. By the user independently of the pod (i.e. with something like my-volume.yaml). In order to create 1 pod with a volume, the user needs to create 2 yaml files and run 2 commands.
2. When a unique volume is specified in a pod or PV spec.
2. Directly in volume definition in the pod/PV object
6. When specifying a volume as part of the pod or PV spec, users have the option to include an extra attribute, e.g. ssTimes, to denote the snapshot schedule.
7. Pros
4. Easy for users to implement and understand
8. Cons
5. The same underlying PD may be used by different pods. In this case, we need to resolve when and how often to take snapshots. If two pods specify the same snapshot time for the same PD, we should not perform two snapshots at that time. However, there is no unique global identifier for a volume defined in a pod definition--its identifying details are particular to the volume plugin used.
6. Replica sets have the same pod spec and support needs to be added so that underlying volume used does not create new snapshots for each member of the set.
3. Only in PV object
9. When specifying a volume as part of the PV spec, users have the option to include an extra attribute, e.g. ssTimes, to denote the snapshot schedule.
10. Pros
7. Slightly cleaner than (b). It logically makes more sense to specify snapshotting at the time of the persistent volume definition (as opposed to in the pod definition) since the snapshot schedule is a volume property.
11. Cons
8. No support for direct volumes
9. Only useful for PVs that do not already have automatic snapshotting tools (e.g. Schedule Snapshot Wizard for iSCSI) -- many do and the same can be achieved with a simple cron job
10. Same problems as (b) with respect to non-unique resources. We may have 2 PV API objects for the same underlying disk and need to resolve conflicting/duplicated schedules.
4. Annotations: key value pairs on API object
12. User experience is the same as (b)
13. Instead of storing the snapshot attribute on the pod/PV API object, save this information in an annotation. For instance, if we define a pod with two volumes we might have {"ssTimes-vol1": [1,5], “ssTimes-vol2”: [2,17]} where the values are slices of integer values representing UTC hours.
14. Pros
11. Less invasive to the codebase than (a-c)
15. Cons
12. Same problems as (b-c) with non-unique resources. The only difference here is the API object representation.
-
Business logic:
-
Does this go on the master, node, or both?
-
Where the snapshot is stored
-
GCE, Amazon: cloud storage
-
Others stored on volume itself (gluster) or external drive (iSCSI)
-
-
Requirements for snapshot operation
- Application flush, sync, and fsfreeze before creating snapshot
-
-
Suggestion:
-
New SnapshotController on master
-
Controller keeps a list of active pods/volumes, schedule for each, last snapshot
-
If controller restarts and we miss a snapshot in the process, just skip it
- Alternatively, try creating the snapshot up to the time + retryPeriod (see 5)
-
If snapshotting call fails, retry for an amount of time specified in retryPeriod
-
Timekeeping mechanism: something similar to cron; keep list of snapshot times, calculate time until next snapshot, and sleep for that period
-
-
Logic to prepare the disk for snapshotting on node
- Application I/Os need to be flushed and the filesystem should be frozen before snapshotting (on GCE PD)
-
-
Alternatives: login entirely on node
-
Problems:
-
If pod moves from one node to another
-
A different node is in now in charge of snapshotting
-
If the volume plugin requires external memory for snapshots, we need to move the existing data
-
-
If the same pod exists on two different nodes, which node is in charge
-
-
-
-
Volume plugin interface/internal API:
-
Allow VolumePlugins to implement the SnapshottableVolumePlugin interface (structure similar to AttachableVolumePlugin)
-
When logic is triggered for a snapshot by the SnapshotController, the SnapshottableVolumePlugin calls out to volume plugin API to create snapshot
-
Similar to volume.attach call
-
-
Other questions:
-
Snapshot period
-
Time or period
-
What is our SLO around time accuracy?
- Best effort, but no guarantees (depends on time or period) -- if going with time.
-
What if we miss a snapshot?
- We will retry (assuming this means that we failed) -- take at the nearest next opportunity
-
Will we know when an operation has failed? How do we report that?
- Get response from volume plugin API, log in kubelet log, generate Kube event in success and failure cases
-
Will we be responsible for GCing old snapshots?
- Maybe this can be explicit non-goal, in the future can automate garbage collection
-
If the pod dies do we continue creating snapshots?
-
How to communicate errors (PD doesn’t support snapshotting, time period unsupported)
-
Off schedule snapshotting like before an application upgrade
-
We may want to take snapshots of encrypted disks. For instance, for GCE PDs, the encryption key must be passed to gcloud to snapshot an encrypted disk. Should Kubernetes handle this?
-
Options, pros, cons, suggestion/recommendation
Example 1b
During pod creation, a user can specify a pod definition in a yaml file. As part of this specification, users should be able to denote a [list of] times at which an existing snapshot command can be executed on the pod’s associated volume.
For a simple example, take the definition of a pod using a GCE PD:
apiVersion: v1 kind: Pod metadata: name: test-pd spec: containers:
- image: gcr.io/google_containers/test-webserver
name: test-container
volumeMounts:
- mountPath: /test-pd name: test-volume volumes:
- name: test-volume gcePersistentDisk: pdName: my-data-disk fsType: ext4
Introduce a new field into the volume spec:
apiVersion: v1 kind: Pod metadata: name: test-pd spec: containers:
- image: gcr.io/google_containers/test-webserver
name: test-container
volumeMounts:
- mountPath: /test-pd name: test-volume volumes:
- name: test-volume gcePersistentDisk: pdName: my-data-disk fsType: ext4
** ssTimes: **[1, 5]
Caveats
-
Snapshotting should not be exposed to the user through the Kubernetes API (via an operation such as create-snapshot) because
-
this does not provide value to the user and only adds an extra layer of indirection/complexity.
-
?
-
Dependencies
-
Kubernetes
-
Persistent volume snapshot support through API