Skip to content

Latest commit

 

History

History
658 lines (503 loc) · 38.2 KB

sap-design-considerations-storage.md

File metadata and controls

658 lines (503 loc) · 38.2 KB
copyright lastupdated keywords subcollection
years
2020, 2024
2024-12-11
SAP, {{site.data.keyword.cloud_notm}} SAP-Certified Infrastructure, {{site.data.keyword.ibm_cloud_sap}}, SAP Workloads
sap

{:shortdesc: .shortdesc} {:codeblock: .codeblock} {:screen: .screen} {:external: target="_blank" .external} {:pre: .pre} {:table: .aria-labeledby="caption"} {:note: .note} {:tip: .tip} {:important: .important}

Storage design considerations

{: #storage-design-considerations}

The SAP systems in a landscape have specific requirements for servers, operating systems, network setup, and supported storage. {: shortdesc}

For SAP workloads that use a Cloud Service Provider, Infrastructure-as-a-Service is similar to existing practices used to run SAP workloads in external data centers or by a data center provider. An SAP landscape has specific requirements for connectivity, between hosts within Cloud IaaS and to external systems. {{site.data.keyword.ibm_cloud_sap}} provides a rich set of functions to improve your SAP landscape, beyond hosting SAP systems.

To assist your project's planning phase, the below sections provide {{site.data.keyword.ibm_cloud_sap}} portfolio design considerations for Storage.

Preface: units of measure for data/information

{: #storage-units-of-measure}

Storage performance refers to the read/write performance from storage file system. Often the throughput Network Storage is shown in Mbps or Gbps, whereas for Local Disk storage is shown in MiB/s.

It is important to note, that Mb (Megabits) is a decimal prefix and MiB (Mebibyte) is a binary prefix so they are on different scales. More confusion arises because MiB (Mebibyte) was commonly known in Microsoft Windows as Megabyte.

For future reference throughout the storage documentation, Mb (Megabits) and MiB (Mebibyte) is used based on the system of units (SI) defined by IEC and adopted by IEEE, ISO, and NIST.

For example:

  • 100 Mbps (Megabits per second), would be 12 MiB/s (Mebibyte per second)
  • 1000 Mbps (Megabits per second) also known as 1 Gbps (Gigabits per second), would be 120 MiB/s (Mebibyte per second)
  • 10 Gbps (Gigabits per second), would be 1200 MiB/s (Mebibyte per second)

Storage configuration for SAP HANA

{: #storage-config-hana}

With any SAP HANA certified profile that is listed as Appliance, storage is already provided or must be attached precisely as described.

When you provision more storage for an SAP HANA instance, you must adhere to mandatory TDI storage requirements. For more information, see IBM System Storage Architecture and Configuration Guide for SAP HANA TDI v2.31.pdf{: external}.

The requirements include multiple volumes that are assigned to the DATA and LOG LVMs, with the striping and multipath enhancements increase I/O performance. For more information, see the following documents:

Storage performance considerations

{: #storage-performance}

It is important to calculate your project requirements before you decide on a storage solution. This calculation is critical for selecting network storage because of storage variations and performance considerations.

Storage impacts on Recovery Time Objective (RTO) of SAP HANA backups

{: #storage-performance-backup-rto}

If you need to restore an SAP HANA system, then the IOPS of your storage has a significant influence on your restore window. Backup windows are not as critical with SAP HANA since all backups are online backups no matter how you configure SAP HANA.

For example, by using {{site.data.keyword.cloud_notm}} {{site.data.keyword.blockstorageshort}}, you can calculate for an approximate 12 TB restore of SAP HANA at a maximum speed. You must create three physical storage devices (block storage iSCSI LUNs) because the maximum size per device is 4 TB. You can create a stripe over these three devices with the Linux® Logical Volume Manager and create one logical device of 12 TB.

The 12 TB provides 3x10 IOPS/GB, which is a total of 122,880 IOPS/GB at 16 KB. This amount gives you a restore time of 1.875 GB per second, or a total restore time of under 2 hours. Since the measurement for the IOPS is taken at a 50/50 distribution of read and write, you can consider the numbers as a lower boundary of restore performance. It is advisable to do backup and restore tests if you rely on a certain restore window.

Network Block Storage considerations

{: #storage-performance-network-block}

The following sections describe the storage considerations that use network block storage for SAP workload scenarios that use various IBM Cloud infrastructure options.

Network Block or File storage for VMware on Classic Infrastructure

{: #storage-performance-network-block-vmware}

Using VMware for SAP workloads on {{site.data.keyword.cloud_notm}} is certified. However it requires choice of the storage and would use the "TDI" delivery model for which you would need to run validation checks to gain SAP Support. Therefore, it is important to consider the correct storage for your VMware hosts when they are running SAP workloads.

For VMware clusters, where SAP workloads are run across multiple VMware vSphere hypervisor nodes, storage must be shared across these hypervisor nodes.

VMware is available to work with Block storage or File storage from {{site.data.keyword.cloud_notm}}. To help you select Block storage or File storage for running SAP on VMware, see VMware Technical Paper on Storage Protocol Comparison{: external}.

When you are using Network Block or File storage, do not expect that certification performance benchmarks to remain the same. Particularly after factoring in the hypervisor overheads as described in Compute Profiles of SAP-certified VMware on Classic Infrastructure.

For VMware datastores (where the virtual machine .VMDK virtual disks are located), the recommendations are:

  • For SAP HANA, use Local SDD disks for the datastore in a RAID10 configuration
  • For SAP HANA, with network storage, use 10 IOPS/GB with each vSphere node that hosts SAP that uses a network interface card with 10 Gbps connection
  • For SAP NetWeaver or SAP AnyDB, with network storage, use at least 4 IOPS/GB with each vSphere node that hosts SAP that uses a network interface card with 10 Gbps connection

To achieve maximum IOPS on a volume, adequate network resources need to be in place. Other considerations include private network usage outside of storage and host side, and application-specific tunings (for example, IP stacks and queue depths). For more information, see Getting started with Block Storage and Getting started with File Storage for more information on storage tiers and performance.

The storage to use with either the VMware manual setup (Bare Metal with VMware OS Image) or VMware automated setup ({{site.data.keyword.cloud_notm}} for VMware Solutions Dedicated), is described in:

Block Storage for Virtual Servers on VPC Infrastructure

{: #storage-performance-network-block-vs-vpc}

For network storage, IOPS per GB is limited and performance depends on the workload. For relational database management systems (RDBMS), it might be advisable to use the same volume for both the database's log and data storage. This setup depends on the behavior of your application.

In general, for typical RDBMS-based applications, a 5 IOPS/GB profile is reasonable.

If your application uses dedicated key performance indicators (KPIs) on storage performance, test the storage throughput before you begin software deployment. By using volume manager-based software RAID (like LVM), you meet almost every KPI.

Network Block Storage for IBM Power Virtual Servers

{: #storage-performance-network-block-power}

For SAP HANA, Tier 1 (NVMe) high performance is required.

For SAP NetWeaver and SAP AnyDB databases (such as IBM Db2 or Oracle DB), Tier 1 (NVMe) is recommended but Tier 3 (SSD) can be used.

Given {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}}s are configured to enterprise-grade standards using Fibre Channel, there are no additional performance considerations.

Sample storage configurations on Classic Infrastructure

{: #sample-classic}

The following sections demonstrate storage configurations in various different SAP workload scenarios, which are using Classic Infrastructure.

Sample storage configuration for IBM Db2 that use Intel Bare Metal

{: #sample-classic-db2}

Table 1 is a sample storage configuration for a 256 GB server with 50,000 SAPS, 1.5 TB at 6,000 IOPS for a central system with SAP. The system uses an IBM Db2 database with external {{site.data.keyword.cloud_notm}} {{site.data.keyword.blockstorageshort}} or {{site.data.keyword.cloud_notm}} {{site.data.keyword.filestorage_short}} (4 IOPS/GB). The calculation for the IOPS is:

  • 6,000 IOPS/1,500 GB = 4 IOPS/GB needed for external storage. It is assumed that 3,000 GB is for backup at 2 IOPS/GB (medium performance.
File system number of volumes Storage type IOPS/GB GB IOPS
/ 1 Internal N/A 150 GB N/A
/boot 1 Internal N/A 0.25 GB N/A
swap 1 Internal N/A 256 GB N/A
/db2 (including logs) 1 Internal N/A 250 GB N/A
sapdata 1 External 4 IOPS/GB 1,500 GB 6,000
backup/log and backup 1 External 2 IOPS/GB 3,000 GB 6,000
{: caption="Sample storage layout based on IOPS calculation" caption-side="top"}

Sample storage configurations on VPC Infrastructure

{: #sample-vpc}

The following sections demonstrate storage configurations in various different SAP workload scenarios, when you are using VPC Infrastructure.

Sample storage configuration for SAP AnyDB with IBM Db2 that uses Intel Virtual Server

{: #sample-intel-vs-vpc-anydb-ibm-db2}

For SAP AnyDB that uses IBM Db2 on mx2-32x256 profile the volumes that are needed are:

  • 1x 500 GB volumes; one block storage volume of 500 GB size, with a custom volume profile that supports up to 10,000 Max IOPS attached to the Virtual Server
  • 1x 2,000 GB volume; one block storage volume of 2,000 GB size, with a lower 4,000 IOPS (medium performance) attached to the Virtual Server for backups

Disk mount points and volumes for IBM Db2

{: #sample-intel-vs-vpc-anydb-ibm-db2-mountpoints}

After you attach the two data volumes, two new virtual disks will appear in the Virtual Server, see the following table. In this example, those disks are vdd, vde, and vdf.

File system Volume Storage type IOPS/GB GB IOPS
/ vdal Pre-configured boot volume N/A 100 GB 3,000
/boot vda2 Pre-configured boot volume N/A 0.25 GB 3,000
/db2 vdd (can vary) Data volume 20 IOPS/GB 500 GB 10,000
backup/log and backup vde (can vary) Data volume 5 IOPS/GB 2,000 GB 4,000
{: caption="Sample storage configuration" caption-side="top"}

Table 1 shows a basic layout of the file system to support an IBM Db2 installation. Generally, an IBM Db2 installation uses subdirectories that can be segmented into independent volumes.

For example, "/db2/<DBSID>", "/db2/<DBSID>/log_dir", and several "sapdata<n>", where the folder "log_dir" contains the online logs files of the database and the "sapdata<n>" contains the data itself. For example, see the Db2 documentation here: Required File Systems for IBM Db2 for Linux, UNIX, and Windows{: external}.

Sample storage configurations for SAP HANA

{: #hana-vs-storage-config}

Further information about storage specifications for Virtual Server are available, the below shows only the configuration steps required.

mx2-8x64, mx2-16x128 and mx2-32x256 profiles

{: #hana-iaas-mx2-16x128-32x256-configure}

The mx2-8x64 profile is certified for SAP Business One on HANA only. {: note}

For Virtual Server created based on the mx2-8x64, mx2-16x128 and mx2-32x256 profiles, there are:

  • 3x 500 GB volumes; three block storage volumes of 500 GB size, with a custom volume profile that supports up to 10,000 Max IOPS attached to the Virtual Server
  • 1x 2,000 GB volume; one block storage volume of 2,000 GB size, with a lower 4,000 IOPS (medium performance) attached to the Virtual Server for backups

After attaching the three data volumes, three new virtual disks will appear in the virtual server, see the table that follows. In this example, those disks are vdd, vde and vdf.

The disks are visible in the operating system of the virtual server as follows:

[root@hana256-vsi ~]# fdisk -l

Disk /dev/vdd: 536.9 GB, 536870912000 bytes, 1048576000 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/vde: 536.9 GB, 536870912000 bytes, 1048576000 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/vdf: 536.9 GB, 536870912000 bytes, 1048576000 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes

These three disks must be managed under the Linux® Logical Volume Manager (LVM), and deployed as logical volumes. In order to achieve that, first put the three devices under LVM control. For example, make them physical volumes:

[root@hana256-vsi ~]# pvcreate /dev/vdd /dev/vde /dev/vdf

Then, create a volume group from the physical volumes. The name of the volume group can be chosen according to your preferences, in our sample it is hana_vg:

[root@hana256-vsi ~]# vgcreate hana_vg /dev/vdd /dev/vde /dev/vdf

After creating the volume group, three logical volumes need to be defined on top. These logical volumes reflect the file system size requirements for SAP HANA. The following commands are for a 256 GB virtual server:

[root@hana256-vsi ~]# lvcreate -i 3 -I 64K -L 256GB -n hana_log_lv hana_vg
[root@hana256-vsi ~]# lvcreate -i 3 -I 64K -L 256GB -n hana_shared_lv hana_vg
[root@hana256-vsi ~]# lvcreate -i 3 -I 64K -l 100%FREE -n hana_data_lv hana_vg

For a 128 GB virtual server, in the example above -L 256GB must be replaced by -L 128GB and for 64 GB by -L 64GB accordingly. These commands will not result in the smallest possible file system size, but they create the smallest configuration, which will fulfill the SAP HANA KPIs. Finally, a file system needs to be created on top of each volume group:

[root@hana256-vsi ~]# mkfs.xfs /dev/mapper/hana_vg-hana_log_lv
[root@hana256-vsi ~]# mkfs.xfs /dev/mapper/hana_vg-hana_data_lv
[root@hana256-vsi ~]# mkfs.xfs /dev/mapper/hana_vg-hana_shared_lv

The following entries to /etc/fstab will mount the file systems after their mount points (/hana/data, /hana/log and /hana/shared) have been created:

/dev/mapper/hana_vg-hana_log_lv    /hana/log xfs defaults,swalloc,nobarrier,inode64
/dev/mapper/hana_vg-hana_shared_lv /hana/shared xfs defaults,inode64 0 0
/dev/mapper/hana_vg-hana_data_lv   /hana/data xfs defaults,largeio,swalloc,inode64 0 0

mx2-48x384 profile

{: #hana-iaas-mx2-48x384-configure}

For Virtual Server created based on the mx2-48x384 profile there are:

  • 3x 500 GB volumes; three block storage volumes of 500 GB size, with a custom volume profile that supports up to 10,000 Max IOPS attached to the Virtual Server is required
  • 4x 100 GB volumes; four block storage volumes of 100 GB size, with a custom volume profile that supports up to 6,000 Max IOPS attached to the Virtual Server is required
  • Optional: 1x 2,000 GB volume; one block storage volume of 2,000 GB size, with a lower 4,000 IOPS (medium performance) attached to the Virtual Server for backups

After attaching the seven data volumes, seven new virtual disks will appear in the Virtual Server, see the table that follows. In this example, those disks are vdd, vde, vdf, vdg, vdh, vdi, vdj.

These three disks must be managed under the Linux® Logical Volume Manager (LVM), and deployed as logical volumes. In order to achieve that, first put the three devices under LVM control. For example, make them physical volumes:

[root@hana384-vsi ~]# pvcreate /dev/vd[d,e,f,g,h,i,j]

Then, two different volume groups need to be created:

[root@hana384-vsi ~]# vgcreate hana_vg /dev/vdh /dev/vdi /dev/vdj
[root@hana384-vsi ~]# vgcreate hana_log_vg /dev/vdd /dev/vde /dev/vdf /dev/vdg

Next, three logical volumes need to be defined on top. These logical volumes reflect the file system size requirements for SAP HANA. The following commands are for a 384 GB virtual server:

[root@hana384-vsi ~]# lvcreate -l 100%VG -i 4 -I 64K  -n hana_log_lv hana_log_vg
[root@hana384-vsi ~]# lvcreate -i 3 -L 384G -I 64K -n hana_shared_lv hana_vg
[root@hana384-vsi ~]# lvcreate -i 3 -l 100%FREE  -I 64K -n hana_data_lv hana_vg

Finally, a file system needs to be created on top of each volume group:

[root@hana384-vsi ~]# mkfs.xfs /dev/mapper/hana_log_vg-hana_log_lv
[root@hana384-vsi ~]# mkfs.xfs /dev/mapper/hana_vg-hana_data_lv
[root@hana384-vsi ~]# mkfs.xfs /dev/mapper/hana_vg-hana_shared_lv

The following entries to /etc/fstab mount the file systems after their mount points (/hana/data, /hana/log and /hana/shared) have been created:

/dev/mapper/hana_log_vg-hana_log_lv    /hana/log xfs defaults,swalloc,nobarrier,inode64
/dev/mapper/hana_vg-hana_shared_lv /hana/shared xfs defaults,inode64 0 0
/dev/mapper/hana_vg-hana_data_lv   /hana/data xfs defaults,largeio,swalloc,inode64 0 0

Sample storage configurations on IBM Power Infrastructure

{: #sample-power}

This is a complementary offering from {{site.data.keyword.IBM_notm}} Power Systems, with low latency access to {{site.data.keyword.cloud_notm}} services {: note}

The following sections demonstrate storage configurations in various different SAP workload scenarios.

When you create new storage volumes, keep the following points in mind:

  • After you provision the storage with Affinity, you cannot change it. Carefully plan your storage layer and make sure that your configuration is correct. Affinity prevents issues with new volume discovery on existing virtual servers.
  • When you finish provisioning new volumes, you can toggle bootable and sharing switches.
  • Make volumes that belong to different OS file systems unique sizes. Otherwise, it is difficult to identify the storage volumes within the operating system.

Storage Guidelines for SAP HANA based on Flexible IOPS

{: #sap-fiops-config}

See Benefits of flexible IOPS to learn more about flexible IOPS. {: note}

The storage tier and capacity that is recommended for deployment of SAP S/4HANA on Power Virtual Server is described in this topic. These recommendations are based on the best practices and to meet the minimum performance criteria defined by HCMT. {: shortdesc}

Recommendation for SAP HANA DB size 128 GB - 768 GB

  • Use 4 volumes of Fixed 5,000 IOPS storage for storing log files. Log files are usually up to 512 GB and need the high performance
  • Use 4 volumes of Tier 0 storage for data file system
  • Use 1 volume of Tier 3 storage for shared file system.

Recommendations for SAP HANA DB size 960 GB - 22.5 TB

  • Use 4 volumes of Tier 0 storage for storing log files. Log files are usually up to 512 GB and need the high performance
  • Use 4 volumes of Tier 3 storage for data file system
  • Use 1 volume of Tier 3 storage for shared file system.

An additional Capacity of 150 GB per compute instance for the Operating System + /usr/sap on Tier 3 Storage profile is recommended. For boot volume Tier 3 is sufficient. {: important}

There are no defined IOPS requirements for shared file system {: note}

Sample storage tier and IOPS mapping

{: #tier-fiops-map}

The tables below shows the mapping of minimum IOPS and its storage tier mapping based on the different SAP profiles.

Certified profile Volume (GB) Minimum IOPS Log-storage tier
ush1-4x128 4 x 16 GB 20,000 Fixed 5,000 IOPS
ush1-4x256 4 x 32 GB 20,000 Fixed 5,000 IOPS
ush1-4x384 4 x 48 GB 20,000 Fixed 5,000 IOPS
ush1-4x512 4 x 64 GB 20,000 Fixed 5,000 IOPS
ush1-4x768 4 x 96 GB 20,000 Fixed 5,000 IOPS
umh-4x960 4 x 128 GB     12,800   Tier 0      
umh-6x1440 4 x 128 GB     12,800   Tier 0      
umh-8x1920 4 x 128 GB     12,800   Tier 0      
umh-10x2400 4 x 128 GB     12,800   Tier 0      
umh-12x2880 4 x 128 GB     12,800   Tier 0      
umh-16x3840 4 x 128 GB     12,800   Tier 0      
umh-20x4800 4 x 128 GB     12,800   Tier 0      
umh-22x5280 4 x 128 GB     12,800   Tier 0      
umh-25x6000 4 x 128 GB     12,800   Tier 0      
umh-30x7200 4 x 128 GB     12,800   Tier 0      
umh-35x8400 4 x 128 GB     12,800   Tier 0      
umh-40x9600 4 x 128 GB     12,800   Tier 0      
umh-50x12000 4 x 128 GB     12,800   Tier 0      
umh-60x14400 4 x 128 GB     12,800   Tier 0     
mh1-8x1440 4 x 128 GB     12,800   Tier 0      
mh1-10x1800 4 x 128 GB     12,800   Tier 0      
mh1-12x2160 4 x 128 GB     12,800   Tier 0      
mh1-16x2880 4 x 128 GB     12,800   Tier 0      
mh1-20x3600 4 x 128 GB     12,800   Tier 0      
mh1-22x3960 4 x 128 GB     12,800   Tier 0      
mh1-25x4500 4 x 128 GB     12,800   Tier 0      
mh1-30x5400 4 x 128 GB     12,800   Tier 0      
mh1-35x6300 4 x 128 GB     12,800   Tier 0      
mh1-40x7200 4 x 128 GB     12,800   Tier 0      
mh1-50x9000 4 x 128 GB     12,800   Tier 0      
mh1-60x10800 4 x 128 GB     12,800   Tier 0     
mh1-70x12600 4 x 128 GB     12,800   Tier 0     
mh1-80x14400 4 x 128 GB     12,800   Tier 0     
mh1-90x16200 4 x 128 GB     12,800   Tier 0     
mh1-100x18000 4 x 128 GB     12,800   Tier 0     
mh1-125x22500 4 x 128 GB     12,800   Tier 0     
ch1-60x3000 4 x 128 GB     12,800   Tier 0      
ch1-70x3500 4 x 128 GB     12,800   Tier 0      
ch1-80x4000 4 x 128 GB     12,800   Tier 0      
ch1-100x5000 4 x 128 GB     12,800   Tier 0     
ch1-120x6000 4 x 128 GB     12,800   Tier 0     
ch1-140x7000 4 x 128 GB     12,800   Tier 0     
bh1-16x1600 4 x 128 GB     12,800   Tier 0     
bh1-20x2000 4 x 128 GB     12,800   Tier 0      
bh1-22x2200 4 x 128 GB     12,800   Tier 0      
bh1-25x2500 4 x 128 GB     12,800   Tier 0      
bh1-30x3000 4 x 128 GB     12,800   Tier 0      
bh1-35x3500 4 x 128 GB     12,800   Tier 0      
bh1-40x4000 4 x 128 GB     12,800   Tier 0      
bh1-50x5000 4 x 128 GB     12,800   Tier 0      
bh1-60x6000 4 x 128 GB     12,800   Tier 0      
bh1-70x7000 4 x 128 GB     12,800   Tier 0      
bh1-80x8000 4 x 128 GB     12,800   Tier 0      
bh1-100x10000 4 x 128 GB     12,800   Tier 0     
bh1-120x12000 4 x 128 GB     12,800   Tier 0     
bh1-140x14000 4 x 128 GB     12,800   Tier 0     
{: class="simple-tab-table"}
{: tab-group="recommended storage-tier"}
{: caption="Recommended storage tier and capacity for log data" caption-side="top"}
{: #log}
{: tab-title="Log"}
Certified profile Volume (GB) Minimum IOPS Data-storage tier
ush1-4x128 4 x 77 GB 7,500 Tier 0
ush1-4x256 4 x 77 GB 7,500 Tier 0
ush1-4x384 4 x 115 GB 7,500 Tier 0
ush1-4x512 4 x 154 GB 7,500 Tier 0
ush1-4x768 4 x 230 GB 7,500 Tier 0
umh-4x960 4 x 722 GB     7,500 Tier 3      
umh-6x1440 4 x 720 GB     7,500 Tier 3      
umh-8x1920 4 x 720 GB     7,500 Tier 3      
umh-10x2400 4 x 720 GB     7,500 Tier 3      
umh-12x2880 4 x 864 GB     7,500 Tier 3      
umh-16x3840 4 x 1,152 GB     7,500 Tier 3      
umh-20x4800 4 x 1,440 GB     7,500 Tier 3      
umh-22x5280 4 x 1,584 GB     7,500 Tier 3      
umh-25x6000 4 x 1,800 GB     7,500 Tier 3      
umh-30x7200 4 x 2,160 GB     7,500 Tier 3      
umh-35x8400 4 x 2,520 GB     7,500 Tier 3      
umh-40x9600 4 x 2,880 GB     7,500 Tier 3      
umh-50x12000 4 x 3,600 GB     7,500 Tier 3      
umh-60x14400 4 x 4,320 GB     7,500 Tier 3      
mh1-8x1440 4 x 648 GB     7,500 Tier 3      
mh1-10x1800 4 x 648 GB     7,500 Tier 3      
mh1-12x2160 4 x 648 GB     7,500 Tier 3      
mh1-16x2880 4 x 864 GB     7,500 Tier 3      
mh1-20x3600 4 x 1,080 GB     7,500 Tier 3      
mh1-22x3960 4 x 1,188 GB     7,500 Tier 3      
mh1-25x4500 4 x 1,350 GB     7,500 Tier 3      
mh1-30x5400 4 x 1,620 GB     7,500 Tier 3      
mh1-35x6300 4 x 1,890 GB     7,500 Tier 3      
mh1-40x7200 4 x 2,160 GB     7,500 Tier 3      
mh1-50x9000 4 x 2,700 GB     7,500 Tier 3      
mh1-60x10800 4 x 3,240 GB     7,500 Tier 3      
mh1-70x12600 4 x 3,780 GB     7,500 Tier 3      
mh1-80x14400 4 x 4,320 GB     7,500 Tier 3      
mh1-90x16200 4 x 4,860 GB     7,500 Tier 3      
mh1-100x18000 4 x 5,400 GB     7,500 Tier 3      
mh1-125x22500 4 x 6,750 GB     7,500 Tier 3      
ch1-60x3000 4 x 900 GB     7,500 Tier 3      
ch1-70x3500 4 x 1,050 GB     7,500 Tier 3      
ch1-80x4000 4 x 1,200 GB     7,500 Tier 3      
ch1-100x5000 4 x 1,500 GB     7,500 Tier 3      
ch1-120x6000 4 x 1,800 GB     7,500 Tier 3      
ch1-140x7000 4 x 2,100 GB     7,500 Tier 3      
bh1-16x1600 4 x 660 GB     7,500 Tier 3      
bh1-20x2000 4 x 660 GB     7,500 Tier 3      
bh1-22x2200 4 x 660 GB     7,500 Tier 3      
bh1-25x2500 4 x 750 GB     7,500 Tier 3      
bh1-30x3000 4 x 900 GB     7,500 Tier 3      
bh1-35x3500 4 x 1,050 GB     7,500 Tier 3      
bh1-40x4000 4 x 1,200 GB     7,500 Tier 3      
bh1-50x5000 4 x 1,500 GB     7,500 Tier 3      
bh1-60x6000 4 x 1,800 GB     7,500 Tier 3      
bh1-70x7000 4 x 2,100 GB     7,500 Tier 3      
bh1-80x8000 4 x 2,400 GB     7,500 Tier 3      
bh1-100x10000 4 x 3,000 GB     7,500 Tier 3      
bh1-120x12000 4 x 3,600 GB     7,500 Tier 3      
bh1-140x14000 4 x 4,200 GB     7,500 Tier 3      
{: class="simple-tab-table"}
{: tab-group="recommended storage-tier"}
{: caption="Recommended storage tier and capacity for data" caption-side="top"}
{: #data}
{: tab-title="Data"}
Certified profile Volume (GB) Shared-storage tier
ush1-4x128 1 x 128 GB Tier 3
ush1-4x256 1 x 256 GB Tier 3
ush1-4x384 1 x 384 GB Tier 3
ush1-4x512 1 x 512 GB Tier 3
ush1-4x768 1 x 768 GB Tier 3
umh-4x960 1 x 1,000 GB Tier 3
umh-6x1440 1 x 1,000 GB Tier 3
mh1-8x1440 1 x 1,000 GB Tier 3
bh1-16x1600 1 x 1,000 GB Tier 3
umh-8x1920 1 x 1,000 GB Tier 3
mh1-10x1800 1 x 1,000 GB Tier 3
bh1-20x2000 1 x 1,000 GB Tier 3
umh-10x2400 1 x 1,000 GB Tier 3
mh1-12x2160 1 x 1,000 GB Tier 3
bh1-22x2200 1 x 1,000 GB Tier 3
bh1-25x2500 1 x 1,000 GB Tier 3
umh-12x2880 1 x 1,000 GB Tier 3
mh1-16x2880 1 x 1,000 GB Tier 3
bh1-30x3000 1 x 1,000 GB Tier 3
ch1-60x3000 1 x 1,000 GB Tier 3
umh-16x3840 1 x 1,000 GB Tier 3
mh1-20x3600 1 x 1,000 GB Tier 3
mh1-22x3960 1 x 1,000 GB Tier 3
bh1-35x3500 1 x 1,000 GB Tier 3
ch1-70x3500 1 x 1,000 GB Tier 3
umh-20x4800 1 x 1,000 GB Tier 3
mh1-25x4500 1 x 1,000 GB Tier 3
bh1-40x4000 1 x 1,000 GB Tier 3
ch1-80x4000 1 x 1,000 GB Tier 3
umh-22x5280 1 x 1,000 GB Tier 3
mh1-30x5400 1 x 1,000 GB Tier 3
bh1-50x5000 1 x 1,000 GB Tier 3
ch1-100x5000 1 x 1,000 GB Tier 3
umh-25x6000 1 x 1,000 GB Tier 3
mh1-35x6300 1 x 1,000 GB Tier 3
bh1-60x6000 1 x 1,000 GB Tier 3
ch1-120x6000 1 x 1,000 GB Tier 3
umh-30x7200 1 x 1,000 GB Tier 3
mh1-40x7200 1 x 1,000 GB Tier 3
bh1-70x7000 1 x 1,000 GB Tier 3
ch1-140x7000 1 x 1,000 GB Tier 3
umh-35x8400 1 x 1,000 GB Tier 3
bh1-80x8000 1 x 1,000 GB Tier 3
umh-40x9600 1 x 1,000 GB Tier 3
mh1-50x9000 1 x 1,000 GB Tier 3
bh1-100x10000 1 x 1,000 GB Tier 3
mh1-60x10800 1 x 1,000 GB Tier 3
umh-50x12000 1 x 1,000 GB Tier 3
mh1-70x12600 1 x 1,000 GB Tier 3
bh1-120x12000 1 x 1,000 GB Tier 3
umh-60x14400 1 x 1,000 GB Tier 3
mh1-80x14400 1 x 1,000 GB Tier 3
bh1-140x14000 1 x 1,000 GB Tier 3
mh1-90x16200 1 x 1,000 GB Tier 3
mh1-100x18000 1 x 1,000 GB Tier 3
mh1-125x22500 1 x 1,000 GB Tier 3
{: class="simple-tab-table"}
{: tab-group="recommended storage-tier"}
{: caption="Recommended storage tier and capacity for shared data" caption-side="top"}
{: #shared}
{: tab-title="Shared"}

Pricing

{: #pricing-sap-hana-fiops}

See the table 4 and table 5 for sample pricing calculation based on the storage requirement defined. The prices indicated are for reference and subject to change. Use the cost estimator{: external} to get an estimate based on your business need.

Fixed 5000 IOPS is limited to 200 GB of storage. {: note}

Sample pricing calculation when you need a storage of 1 GB

Storage Required = 1 GB Calculation $Price/Month IOPS Performance
Tier 0 1 GB x $.24 $0.240 25 IOPS
Fixed 5,000 IOPS 1 GB x $.288 $0.288 5,000 IOPS
{: class="simple-table"}
{: caption="Sample pricing when the storage required is 1 GB" caption-side="top"}

Sample pricing calculation when you need a storage of 100 GB

Storage Required = 100 GB Calculation $Price/Month IOPS Performance
Tier 0 100 GB x $.24 $24 2,500 IOPS
Fixed 5,000 IOPS 100 GB x $.288 $29 5,000 IOPS
{: class="simple-table"}
{: caption="Sample pricing when the storage required is 100 GB" caption-side="top"}

Sample storage configuration for SAP NetWeaver that use the IBM Power Virtual Server

{: #sample-power-netweaver}

Table 1 is a sample storage configuration for a {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}} for SAP NetWeaver on Linux.

The storage can't be combined within one {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}}; it must be either Tier 1 or Tier 3. When you provision the Linux server, you automatically get the standard LVM configuration for the root volume group. These file systems are included in table 1. Disk sizes depend on whether the installation is Greenfield, or whether the server is a copy of an on-premises AIX server that you decided to use as a sizing reference.

The naming convention for the LVM entries is optional, but the advice is to include the SID of your SAP NetWeaver system especially if you plan to install one or more instances.

Storage Volume group Logical volume Mount point
OS disk Default configuration Default configuration Default configuration
Application disk app<sid>vg lvusrsap /usr/sap
lvusrsap<SID> /usr/sap/<SID
lvusrsapmnt /sapmnt/<SID>
lvusrsaptrans /usr/sap/trans
lvsapDAH /usr/sap/DAH
{: caption="Sample storage layout for Linux" caption-side="top"}

Sample storage configuration for Oracle DB on IBM AIX that use the IBM Power Virtual Server

{: #ample-power-anydb-oracle-aix}

Table 2 is a sample configuration for an AIX {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}} for an SAP NetWeaver application server that uses Oracle as the example.

The storage cannot be combined within the same {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}}, and can be either Tier 1 or Tier 3. The recommendation is to provision three more disks to enable separation between the OS, database, and application layer. Disk size depends on if the installation is Greenfield or if the server is a copy of an "on-premises" AIX server you decided to use as a sizing reference.

The naming convention for the LVM entries is optional, but the advice is to include the SID of your SAP NetWeaver system, especially if you plan to install one or more instances.

Storage Volume group Logical volume Mount point
OS disk Default configuration Default configuration Default configuration
Application disk app<sid>vg lvusrsap /usr/sap
lvusrsap<SID> /usr/sap/<SID>
lvusrsapmnt /sapmnt/<SID>
lvusrsaptrans /usr/sap/trans
lvsapDAH /usr/sap/DAH
Database storage db<sid>vg lv<SID>arch /oracle/<SID>/oraarch
lv<SID>reorg /oracle/<SID>/sapreorg
lv<SID>origlogA /oracle/<SID>/origlogA
lv<SID>origlogB /oracle/<SID>/origlogA
lv<SID>ora /oracle/<SID>
lv<SID>sapdata1 /oracle/<SID>/sapdata1
lv<SID>sapdata2 /oracle/<SID>/sapdata2
lvorastage /oracle/stage
lv<SID>sapdata3 /oracle/<SID>/sapdata3
lv<SID>sapdata4 /oracle/<SID>/sapdata4
lv<SID>oraclient /oracle/client
{: caption="Sample storage layout for Oracle" caption-side="top"}

For more information, see SAP Note 2172935{: external}.

Sample storage configuration for {{site.data.keyword.Db2_on_Cloud_short}} on IBM AIX that use the IBM Power Virtual Server

{: #sample-power-anydb-ibm-db2-aix}

Table 3 is a sample storage configuration for an AIX {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}} for a {{site.data.keyword.Db2_on_Cloud_short}} server.

The storage cannot be combined within the same {{site.data.keyword.IBM_notm}} {{site.data.keyword.powerSys_notm}}, and can be either Tier 1 or Tier 3. The recommendations are to provision three more disks to enable separation between the OS, database, and application layer. Disk size depends on whether the installation is Greenfield or the server is a copy of an "on-premises" AIX server that you decided to use as a sizing reference.

The naming convention for the LVM entries is optional, but the advice is to include the SID of your SAP NetWeaver system especially if you plan to install one or more instances.

Storage Volume group Logical volume Mount point
OS disk Default configuration Default configuration Default configuration
Application disk app<sid>vg lvusrsap /usr/sap
lvusrsap<SID> /usr/sap/<SID>
lvusrsapmnt /sapmnt/<SID>
lvusrsaptrans /usr/sap/trans
lvsapDAH /usr/sap/DAH
Db2 database storage I <sid>db2vg loglv<SID> NA
lv<SID>db2 /db2/<SID>
lvhome<SID> /db2/db2<SID>
lv<SID>db2dump /db2/<SID>/db2dump
lv<SID>logdir /db2/<SID>/log_dir
lv<SID>log_archive /db2/<SID>/log_archive
lv<SID>saptmp /db2/<SID>/saptemp1
lv<SID>db2sw /db2/db2/<DBSID>/db2_sw
Db2 database storage II <sid>db2datvg lv<SID>sapdata1 /db2/<SID>/sapdata1
``lvsapdata2` /db2/<SID>/sapdata2
``lvsapdata3` /db2/<SID>/sapdata3
``lvsapdata4` /db2/<SID>/sapdata4
{: caption="Sample storage layout for Db2 on Cloud" caption-side="top"}

For more information, see Required File Systems for IBM Db2 for Linux, UNIX, and Windows{: external} and SAP Note 1707361{: external}.