3 Maintaining Oracle Exadata Storage Servers
Oracle Exadata Storage Servers contain disks and memory devices that might require maintenance.
Note:
- All procedures in this chapter are applicable to Oracle Exadata and Oracle Exadata Storage Expansion Rack.
- For ease of reading, the name "Oracle Exadata Rack" is used when information refers to both Oracle Exadata and Oracle Exadata Storage Expansion Rack.
- Maintaining Oracle Exadata Storage Servers
This section describes how to perform maintenance on Oracle Exadata Storage Servers. - Using Exadata Extended (XT) Storage Servers
Oracle Exadata Extended (XT) Storage Server offers a lower cost storage option that can be used for infrequently accessed, older, or regulatory data. - Maintaining the Hard Disks of Oracle Exadata Storage Servers
- Maintaining Flash Disks on Oracle Exadata Storage Servers
- Maintaining PMEM Devices on Oracle Exadata Storage Servers
Persistent memory (PMEM) devices reside in Exadata X8M-2 and X9M-2 storage server models with High Capacity (HC) or Extreme Flash (EF) storage. - Maintaining the M.2 Disks of Oracle Exadata Storage Server
Oracle Exadata X7 and later systems come with two internal M.2 devices that contain the system area. - Resizing Grid Disks
You can resize grid disks and Oracle ASM disk groups to shrink one with excess free space and increase the size of another that is near capacity. - Using the Oracle Exadata System Software Rescue Procedure
In the rare event that the system disks fail simultaneously, you must use the Oracle Exadata Storage Server rescue procedure to recover the system. - Changing Existing Elastic Configurations for Storage Cells
You can modify the capacity of your Oracle Exadata using elastic configuration. - Managing Disk Controller Batteries
This section applies only to Exadata systems prior to X6 that use batteries. Newer systems have CVPM02 (Cache Vault), which is a super cap and not a battery. - Managing F20 PCIe Energy Storage Modules
Sun Flash Accelerator F20 PCIe cards are used in Oracle Exadata X3 models. - Exadata Storage Server LED Indicator Descriptions
The indicator LEDs on Oracle Exadata storage servers help you to verify the system status and identify components that require servicing. - Exadata Storage Server Images
The Exadata Storage Server models have different external layouts and physical appearance.
3.1 Maintaining Oracle Exadata Storage Servers
This section describes how to perform maintenance on Oracle Exadata Storage Servers.
- Shutting Down Exadata Storage Server
When performing maintenance on Exadata Storage Servers, it may be necessary to power down or restart the cell. - Checking Status of a Rebalance Operation
When dropping or adding a disk, you can check the status of the Oracle ASM rebalance operation. - Enabling Network Connectivity with the Diagnostic ISO
If a storage server does not restart, the diagnostic ISO may be needed to access the cell so it can be manually repaired.
Parent topic: Maintaining Oracle Exadata Storage Servers
3.1.1 Shutting Down Exadata Storage Server
When performing maintenance on Exadata Storage Servers, it may be necessary to power down or restart the cell.
If Exadata Storage Server is to be shut down when one or more databases are running, then you must verify that taking Exadata Storage Server offline will not impact Oracle ASM disk group and database availability. The ability to take Exadata Storage Server offline without affecting database availability depends on the level of Oracle ASM redundancy used on the affected disk groups. Availability also depends on the current status of disks in other Exadata Storage Servers that have mirror copies of data for the Exadata Storage Server that you are taking offline.
Parent topic: Maintaining Oracle Exadata Storage Servers
3.1.2 Checking Status of a Rebalance Operation
When dropping or adding a disk, you can check the status of the Oracle ASM rebalance operation.
-
The rebalance operation may have completed successfully. Check the Oracle ASM alert logs to confirm.
-
The rebalance operation may be currently running. Check the
GV$ASM_OPERATION
view to determine if the rebalance operation is still running. -
The rebalance operation may have failed. Check the
V$ASM_OPERATION.ERROR
view to determine if the rebalance operation failed. -
Rebalance operations from multiple disk groups can be done on different Oracle ASM instances in the same cluster if the physical disk being replaced contains Oracle ASM disks from multiple disk groups. One Oracle ASM instance can run one rebalance operation at a time. If all Oracle ASM instances are busy, then rebalance operations are queued.
Note:
For Oracle Exadata Storage Servers running Oracle Exadata System Software release 12.1.2.0 with Oracle Database release 12.1.0.2 with BP4, Oracle ASM sends an e-mail about the status of a rebalance operation. In earlier releases, the administrator had to check the status of the operation.
Parent topic: Maintaining Oracle Exadata Storage Servers
3.1.3 Enabling Network Connectivity with the Diagnostic ISO
If a storage server does not restart, the diagnostic ISO may be needed to access the cell so it can be manually repaired.
The diagnostic ISO should be used after other boot methods, such as using the USB, do not work.
The following procedure enables networking with the diagnostic ISO so files can be transferred to repair the cell:
Parent topic: Maintaining Oracle Exadata Storage Servers
3.2 Using Exadata Extended (XT) Storage Servers
Oracle Exadata Extended (XT) Storage Server offers a lower cost storage option that can be used for infrequently accessed, older, or regulatory data.
- About Oracle Exadata Extended (XT) Storage Servers
Oracle Exadata Extended (XT) Storage Servers help you extend the operational and management benefits of Exadata Database Machine to rarely accessed data that must be kept online. - What Data Can Be Stored on Oracle Exadata Extended (XT) Storage Servers?
Oracle Exadata Extended (XT) Storage Servers are intended provide lower-cost storage for infrequently accessed, older, or regulatory data. - Enabling Smart Scan on Exadata Extended (XT) Storage Servers
If you purchase Oracle Exadata System Software licenses for your Oracle Exadata Extended (XT) Storage Servers, then you can enable features such as Smart Scan and Storage Indexes to improve performance.
Parent topic: Maintaining Oracle Exadata Storage Servers
3.2.1 About Oracle Exadata Extended (XT) Storage Servers
Oracle Exadata Extended (XT) Storage Servers help you extend the operational and management benefits of Exadata Database Machine to rarely accessed data that must be kept online.
Each XT storage server includes twelve high-capacity disk drives. However, to achieve a lower cost, flash devices are not included. Also, Oracle Exadata System Software licensing is optional, with some software features disabled unless licensed. Hybrid Columnar Compression is included without requiring Oracle Exadata System Software licenses.
XT storage servers use the same RDMA Network Fabric as the other servers in your Oracle Exadata Rack. XT storage servers add storage capacity while remaining transparent to applications, transparent to SQL, and retaining the same operational model. You can use the same security model and encryption used for your other Exadata storage servers.
You can add XT storage servers to Oracle Exadata Rack X4 or newer, including Eighth Rack configurations. You must add at least 2 servers initially. After the initial 2 servers, you can add XT storage servers as needed. To implement high redundancy, you must have a minimum of 3 XT storage servers. XT storage servers follow the same placement patterns as High Capacity (HC) and Extreme Flash (EF) storage servers.
An Oracle ASM disk group should use storage provided by only one type of storage server (HC, EF, or XT). After adding the XT storage servers to your rack, create new disk groups to use the storage. The default disk group name for XT storage servers is XTND. However, you can use a different name as required.
XT storage servers provide fully integrated storage for Oracle Database. You can use the new disk group with database features such as Oracle Partitioning, Oracle Automatic Data Optimization, and Oracle Advanced Compression.
Parent topic: Using Exadata Extended (XT) Storage Servers
3.2.2 What Data Can Be Stored on Oracle Exadata Extended (XT) Storage Servers?
Oracle Exadata Extended (XT) Storage Servers are intended provide lower-cost storage for infrequently accessed, older, or regulatory data.
XT storage servers help you to keep all your required data online and available for queries. This includes data such as:
- Historical data
- Images, BLOBs, contracts, and other large table-based objects
- Compliance and regulatory data
- Local backups
The XT storage servers can also provide storage for development databases which have less stringent performance requirements compared to production databases.
Parent topic: Using Exadata Extended (XT) Storage Servers
3.2.3 Enabling Smart Scan on Exadata Extended (XT) Storage Servers
If you purchase Oracle Exadata System Software licenses for your Oracle Exadata Extended (XT) Storage Servers, then you can enable features such as Smart Scan and Storage Indexes to improve performance.
Parent topic: Using Exadata Extended (XT) Storage Servers
3.3 Maintaining the Hard Disks of Oracle Exadata Storage Servers
Every Oracle Exadata Storage Server in Oracle Exadata Rack has a system area, which is where the Oracle Exadata System Software system software resides. In Oracle Exadata X7 and later systems, two internal M.2 devices contain the system area. In all other systems, the first two disks of Oracle Exadata Storage Server are system disks and the portions on these system disks are referred to as the system area.
In Oracle Exadata X7 and later systems, all the hard disks in the cell are data disks. In systems prior to Oracle Exadata X7, the non-system area of the system disks, referred to as data partitions, is used for normal data storage. All other disks in the cell are data disks.
Starting in Oracle Exadata System Software release 11.2.3.2.0, if there is a disk failure, then Oracle Exadata System Software sends an alert stating that the disk can be replaced, and, after all data has been rebalanced out from that disk, turns on the blue OK to Remove LED for the hard disk with predictive failure. In Oracle Exadata System Software releases earlier than 11.2.3.2.0, the amber Fault-Service Required LED was turned on for a hard disk with predictive failure, but not the blue LED. In these cases, it is necessary to manually check if all data has been rebalanced out from the disk before proceeding with disk replacement.
Starting with Oracle Exadata System Software release 18.1.0.0.0 and Oracle Exadata X7 systems, there is an additional white Do Not Service LED that indicates when redundancy is reduced to inform system administrators or field engineers that the storage server should not be powered off for services. When redundancy is restored, Oracle Exadata System Software automatically turns off the Do Not Service LED to indicate that the cell can be powered off for services.
For a hard disk that has failed, both the blue OK to Remove LED and the amber Fault-Service Required LED are turned on for the drive indicating that disk replacement can proceed. The behavior is the same in all releases. The drive LED light is a solid light in Oracle Exadata System Software releases 11.2.3.2.0 and later; the drive LED blinks in earlier releases.
Note:
Oracle Exadata Rack is online and available while replacing the Oracle Exadata Storage Server physical disks.This section contains the following topics:
- Monitoring the Status of Hard Disks
You can monitor the status of a hard disk by checking its attributes with the CellCLILIST PHYSICALDISK
command. - Monitoring Hard Disk Controller Write-through Caching Mode
The hard disk controller on each Oracle Exadata Storage Server periodically performs a discharge and charge of the controller battery. During the operation, the write cache policy changes from write-back caching to write-through caching. - Replacing a Hard Disk Due to Disk Failure
- Replacing a Hard Disk Due to Disk Problems
You may need to replace a hard disk because the disk is inwarning - predictive failure
status. - Replacing a Hard Disk Due to Bad Performance
A single bad hard disk can degrade the performance of other good disks. It is better to remove the bad disk from the system than let it remain. - Replacing a Hard Disk Proactively
Exadata Storage software has a complete set of automated operations for hard disk maintenance, when a hard disk has failed or has been flagged as a problematic disk. But there are situations where a hard disk has to be removed proactively from the configuration. - Moving All Drives to Another Exadata Storage Server
It may necessary to move all drives from one Exadata Storage Server to another Exadata Storage Server. - Repurposing a Hard Disk
You may want to delete all data on a disk, and then use the disk for another purpose. - Removing and Replacing the Same Hard Disk
What happens if you accidentally remove the wrong hard disk? - Re-Enabling a Hard Disk That Was Rejected
If a physical disk was rejected because it was inserted into the wrong slot, you can re-enable the disk.
Related Topics
Parent topic: Maintaining Oracle Exadata Storage Servers
3.3.1 Monitoring the Status of Hard Disks
You can monitor the status of a hard disk by checking its attributes with the CellCLI LIST PHYSICALDISK
command.
For example, a hard disk status equal to failed
(the status for failed hard disks was critical
in earlier releases), or warning - predictive failure
is probably having problems and needs to be replaced. The disk firmware maintains the error counters, and marks a drive with Predictive Failure
when internal thresholds are exceeded. The drive, not the cell software, determines if it needs replacement.
When disk I/O errors occur, Oracle ASM performs bad extent repair for read errors due to media errors. The disks will stay online, and no alerts are sent. When Oracle ASM gets a read error on a physically-addressed metadata block, it does not have mirroring for the blocks, and takes the disk offline. Oracle ASM then drops the disk using the FORCE
option.
The Oracle Exadata Storage Server hard disk statuses are as follows:
-
Oracle Exadata System Software release 11.2.3.3 and later:
- normal
- normal - dropped for replacement
- normal - confinedOnline
- normal - confinedOnline - dropped for replacement
- not present
- failed
- failed - dropped for replacement
- failed - rejected due to incorrect disk model
- failed - rejected due to incorrect disk model - dropped for replacement
- failed - rejected due to wrong slot
- failed - rejected due to wrong slot - dropped for replacement
- warning - confinedOnline
- warning - confinedOnline - dropped for replacement
- warning - peer failure
- warning - poor performance
- warning - poor performance - dropped for replacement
- warning - poor performance, write-through caching
- warning - predictive failure, poor performance
- warning - predictive failure, poor performance - dropped for replacement
- warning - predictive failure, write-through caching
- warning - predictive failure
- warning - predictive failure - dropped for replacement
- warning - predictive failure, poor performance, write-through caching
- warning - write-through caching
-
Oracle Exadata System Software release 11.2.3.2:
- normal
- normal - confinedOnline
- not present
- failed
- failed - rejected due to incorrect disk model
- failed - rejected due to wrong slot
- warning - confinedOnline
- warning - peer failure
- warning - poor performance
- warning - poor performance, write-through caching
- warning - predictive failure, poor performance
- warning - predictive failure, write-through caching
- warning - predictive failure
- warning - predictive failure, poor performance, write-through caching
- warning - write-through caching
-
Oracle Exadata System Software release 11.2.3.1.1 and earlier:
- normal
- critical
- poor performance
- predictive failure
- not present
Related Topics
3.3.2 Monitoring Hard Disk Controller Write-through Caching Mode
The hard disk controller on each Oracle Exadata Storage Server periodically performs a discharge and charge of the controller battery. During the operation, the write cache policy changes from write-back caching to write-through caching.
Write-through cache mode is slower than write-back cache mode. However, write-back cache mode has a risk of data loss if the Oracle Exadata Storage Server loses power or fails. For Oracle Exadata System Software releases earlier than release 11.2.1.3, the operation occurs every month. For Oracle Exadata System Software release 11.2.1.3.0 and later, the operation occurs every three months, for example, at 01:00 on the 17th day of January, April, July and October.
3.3.3 Replacing a Hard Disk Due to Disk Failure
A hard disk outage can cause a reduction in performance and data redundancy. Therefore, the disk should be replaced with a new disk as soon as possible. When the disk fails, the Oracle ASM disks associated with the grid disks on the hard disk are automatically dropped with the FORCE
option, and an Oracle ASM rebalance follows to restore the data redundancy.
An Exadata alert is generated when a disk fails. The alert includes specific instructions for replacing the disk. If you have configured the system for alert notifications, then the alert is sent by e-mail to the designated address.
After the hard disk is replaced, the grid disks and cell disks that existed on the previous disk in that slot are re-created on the new hard disk. If those grid disks were part of an Oracle ASM group, then they are added back to the disk group, and the data is rebalanced on them, based on the disk group redundancy and ASM_POWER_LIMIT
parameter.
Note:
For storage servers running Oracle Exadata System Software release 12.1.2.0 with Oracle Database release 12.1.0.2 with BP4, Oracle ASM sends an e-mail about the status of a rebalance operation. In earlier releases, the administrator had to check the status of the operation.
For earlier releases, check the rebalance operation status as described in Checking Status of a Rebalance Operation.
The following procedure describes how to replace a hard disk due to disk failure:
In rare cases, the automatic firmware update may not work, and the LUN is not rebuilt. This can be confirmed by checking the ms-odl.trc
file.
See Also:
-
Oracle Database Reference for information about the
V$ASM_OPERATION
view -
Oracle Automatic Storage Management Administrator's Guide for information about the rebalance operation
3.3.4 Replacing a Hard Disk Due to Disk Problems
You may need to replace a hard disk because the disk is in warning - predictive failure
status.
The predictive failure
status indicates that the hard disk will soon fail, and should be replaced at the earliest opportunity. The Oracle ASM disks associated with the grid disks on the hard drive are automatically dropped, and an Oracle ASM rebalance relocates the data from the predictively failed disk to other disks.
If the drop did not complete before the hard drive dies, then refer to Replacing a Hard Disk Due to Disk Failure.
An alert is sent when the disk is removed. After replacing the hard disk, the grid disks and cell disks that existed on the previous disk in the slot are re-created on the new hard disk. If those grid disks were part of an Oracle ASM disk group, then they are added back to the disk group, and the data is rebalanced based on disk group redundancy and the ASM_POWER_LIMIT
parameter.
Note:
On Oracle Exadata Storage Servers running Oracle Exadata System Software release 12.1.2.0 with Oracle Database release 12.1.0.2 with BP4, Oracle ASM sends an e-mail about the status of a rebalance operation. In earlier releases, the administrator had to check the status of the operation.
For earlier releases, check the rebalance operation status as described in Checking Status of a Rebalance Operation.
See Also:
- Oracle Database
Reference for information about the
V$ASM_OPERATION
view - Oracle Automatic Storage Management Administrator's Guide for information about tuning rebalance operations.
- Oracle Database
Reference for information about querying the
V$ASM_DISK_STAT
view ALTER CELL VALIDATE CONFIGURATION
command in Oracle Exadata System Software User's Guide
3.3.5 Replacing a Hard Disk Due to Bad Performance
A single bad hard disk can degrade the performance of other good disks. It is better to remove the bad disk from the system than let it remain.
Starting with Oracle Exadata System Software release 11.2.3.2, an underperforming disk is automatically identified and removed from active configuration. Oracle Exadata Database Machine then runs a set of performance tests. When poor disk performance is detected by CELLSRV, the cell disk status changes to normal - confinedOnline
, and the hard disk status changes to warning - confinedOnline
.
The following conditions trigger disk confinement:
-
Disk stopped responding. The cause code in the storage alert log is
CD_PERF_HANG
. -
Slow cell disk such as the following:
-
High service time threshold (cause code
CD_PERF_SLOW_ABS
) -
High relative service time threshold (cause code
CD_PERF_SLOW_RLTV
)
-
-
High read or write latency such as the following:
-
High latency on writes (cause code
CD_PERF_SLOW_LAT_WT
) -
High latency on reads (cause code
CD_PERF_SLOW_LAT_RD
) -
High latency on reads and writes (cause code
CD_PERF_SLOW_LAT_RW
) -
Very high absolute latency on individual I/Os happening frequently (cause code
CD_PERF_SLOW_LAT_ERR
)
-
-
Errors such as I/O errors (cause code
CD_PERF_IOERR
).
If the disk problem is temporary and passes the tests, then it is brought back into the configuration. If the disk does not pass the tests, then it is marked as poor performance
, and Oracle Auto Service Request (ASR) submits a service request to replace the disk. If possible, Oracle ASM takes the grid disks offline for testing. If Oracle ASM cannot take the disks offline, then the cell disk status stays at normal - confinedOnline
until the disks can be taken offline safely.
The disk status change is associated with the following entry in the cell alert history:
MESSAGE ID date_time info "Hard disk entered confinement status. The LUN
n_m changed status to warning - confinedOnline. CellDisk changed status to normal
- confinedOnline. Status: WARNING - CONFINEDONLINE Manufacturer: name Model
Number: model Size: size Serial Number: serial_number Firmware: fw_release
Slot Number: m Cell Disk: cell_disk_name Grid Disk: grid disk 1, grid disk 2
... Reason for confinement: threshold for service time exceeded"
The following would be logged in the storage cell alert log:
CDHS: Mark cd health state change cell_disk_name with newState HEALTH_BAD_
ONLINE pending HEALTH_BAD_ONLINE ongoing INVALID cur HEALTH_GOOD
Celldisk entering CONFINE ACTIVE state with cause CD_PERF_SLOW_ABS activeForced: 0
inactiveForced: 0 trigger HistoryFail: 0, forceTestOutcome: 0 testFail: 0
global conf related state: numHDsConf: 1 numFDsConf: 0 numHDsHung: 0 numFDsHung: 0
...
Note:
In releases earlier than Oracle Exadata System Software release 11.2.3.2, use the CALIBRATE
command to identify a bad hard disk, and look for very low throughput and IOPS for each hard disk.
The following procedure describes how to remove a hard disk once the bad disk has been identified:
See Also:
-
Oracle Automatic Storage Management Administrator's Guide for information about dropping a disk from a disk group
3.3.6 Replacing a Hard Disk Proactively
Exadata Storage software has a complete set of automated operations for hard disk maintenance, when a hard disk has failed or has been flagged as a problematic disk. But there are situations where a hard disk has to be removed proactively from the configuration.
In the CellCLI ALTER PHYSICALDISK
command, the DROP FOR
REPLACEMENT
option checks if a normal functioning hard disk can be
removed safely without the risk of data loss. However, after the execution of the
command, the grid disks on the hard disk are inactivated on the storage cell and set
to offline in the Oracle ASM disk groups.
To reduce the risk of having a disk group without full redundancy and proactively replace a hard disk, follow this procedure:
3.3.7 Moving All Drives to Another Exadata Storage Server
It may necessary to move all drives from one Exadata Storage Server to another Exadata Storage Server.
This need may occur when there is a chassis-level component failure, such as a motherboard or ILOM failure, or when troubleshooting a hardware problem.
See Also:
-
Oracle Auto Service Request Quick Installation Guide for Oracle Exadata Database Machine
- for information about the CellCLI utility
-
Oracle Automatic Storage Management Administrator's Guide for information about disk repair time
3.3.8 Repurposing a Hard Disk
You may want to delete all data on a disk, and then use the disk for another purpose.
Before repurposing a hard disk, ensure that you have copies of the data that is on the disk.
If you use this procedure for the system disks (disk 0 and disk1), then only the data partitions are erased, not the system partitions.
Related Topics
3.3.9 Removing and Replacing the Same Hard Disk
What happens if you accidentally remove the wrong hard disk?
If you inadvertently remove the wrong hard disk, then put the disk back. It will automatically be added back in the Oracle ASM disk group, and its data is resynchronized.
Note:
When replacing disk due to disk failure or disk problems, the LED is lit on the disk for identification.
3.3.10 Re-Enabling a Hard Disk That Was Rejected
If a physical disk was rejected because it was inserted into the wrong slot, you can re-enable the disk.
Run the following command:
Caution:
The following command removes all data on the physical disk.
CellCLI> ALTER PHYSICALDISK hard_disk_name reenable force
The following is an example of the output from the command:
Physical disk 20:0 was reenabled.
3.4 Maintaining Flash Disks on Oracle Exadata Storage Servers
Data is mirrored across Exadata Cells, and write operations are sent to at least two storage cells. If a flash card in one Oracle Exadata Storage Server has problems, then the read and write operations are serviced by the mirrored data in another Oracle Exadata Storage Server. No interruption of service occurs for the application.
If a flash card fails while in write-back mode, then Oracle Exadata System Software determines the data in the flash cache by reading the
data from the mirrored copy. The data is then written to the cell that had the failed
flash card. The location of the data lost in the failed flash cache is saved by Oracle Exadata System Software at the time of the flash failure.
Resilvering then starts by replacing the lost data with the mirrored copy. During
resilvering, the grid disk status is ACTIVE -- RESILVERING WORKING
. If
the cache is in write-through mode, then the data in the failed device is already
present on the data grid disk, so there is no need for resilvering.
- Replacing a Flash Disk Due to Flash Disk Failure
Each Oracle Exadata Storage Server is equipped with flash devices. - About Flash Disk Degraded Performance Statuses
If a flash disk has degraded performance, you might need to replace the disk. - Replacing a Flash Disk Due to Flash Disk Problems
- Performing a Hot Pluggable Replacement of a Flash Disk
Starting with Oracle Exadata Database Machine X7, flash disks are hot-pluggable on both Extreme Flash (EF) and High Capacity (HC) storage servers. - Enabling Write Back Flash Cache
Write operations serviced by flash instead of by disk are referred to as write-back flash cache. - Disabling Write Back Flash Cache
You can disable the Write-Back Flash Cache by enabling Write-Through Flash Cache. - Enabling Flash Cache Compression
- Monitoring Exadata Smart Flash Cache Usage Statistics
- Disabling Flash Cache Compression
Parent topic: Maintaining Oracle Exadata Storage Servers
3.4.1 Replacing a Flash Disk Due to Flash Disk Failure
Each Oracle Exadata Storage Server is equipped with flash devices.
Starting with Oracle Exadata Database Machine X7, the
flash devices are hot-pluggable on the Oracle Exadata Storage Servers. When performing a hot-pluggable replacement of a flash device on
Oracle Exadata Storage Servers for X7 or later, the
disk status should be Dropped for replacement
, and the power LED on
the flash card should be off, which indicates the flash disk is ready for online
replacement.
Caution:
Removing a card with power LED on could result in a system crash. If a failed disk has a status ofFailed – dropped for replacement
but the power LED is still on, contact Oracle Support Services.
For Oracle Exadata Database Machine X6 and earlier, the flash devices are hot-pluggable on Extreme Flash (EF) storage servers, but not on High Capacity (HC) storage servers. On HC storage servers, you need to power down the storage servers before replacing them.
To identify a failed flash disk, use the following command:
CellCLI> LIST PHYSICALDISK WHERE disktype=flashdisk AND status=failed DETAIL
The following is an example of the output from an Extreme Flash storage server:
name: NVME_10
deviceName: /dev/nvme7n1
diskType: FlashDisk
luns: 0_10
makeModel: "Oracle NVMe SSD"
physicalFirmware: 8DV1RA13
physicalInsertTime: 2016-09-28T11:29:13-07:00
physicalSerial: CVMD426500E21P6LGN
physicalSize: 1.4554837569594383T
slotNumber: 10
status: failed
The following is an example of the output from an Oracle Flash Accelerator F160 PCIe Card:
CellCLI> LIST PHYSICALDISK WHERE DISKTYPE=flashdisk AND STATUS=failed DETAIL
name: FLASH_5_1
deviceName: /dev/nvme1n1
diskType: FlashDisk
luns: 5_1
makeModel: "Oracle Flash Accelerator F160 PCIe Card"
physicalFirmware: 8DV1RA13
physicalInsertTime: 2016-11-30T21:24:45-08:00
physicalSerial: 1030M03UYM
physicalSize: 1.4554837569594383T
slotNumber: "PCI Slot: 5; FDOM: 1"
status: failed
The following is an example of the output from a Sun Flash Accelerator F40 PCIe card:
name: FLASH_5_3
diskType: FlashDisk
luns: 5_3
makeModel: "Sun Flash Accelerator F40 PCIe Card"
physicalFirmware: TI35
physicalInsertTime: 2012-07-13T15:40:59-07:00
physicalSerial: 5L002X4P
physicalSize: 93.13225793838501G
slotNumber: "PCI Slot: 5; FDOM: 3"
status: failed
For the PCIe cards, the name
and slotNumber
attributes show the PCI slot and the FDOM number. For Extreme Flash storage servers,
the slotNumber
attribute shows the NVMe slot on the front panel.
On Oracle Exadata Database Machine X7 and later systems, all flash disks are in the form of an Add-in-Card (AIC), which is inserted into a PCIe slot on the motherboard. The slotNumber
attribute shows the PCI number and FDOM number, regardless of whether it is an EF or HC storage server.
If an flash disk is detected to have failed, then an alert is generated indicating that the flash disk, as well as the LUN on it, has failed. The alert message includes either the PCI slot number and FDOM number or the NVMe slot number. These numbers uniquely identify the field replaceable unit (FRU). If you have configured the system for alert notification, then an alert is sent by e-mail message to the designated address.
A flash disk outage can cause reduction in performance and data redundancy. The failed disk should be replaced with a new flash disk at the earliest opportunity. If the flash disk is used for flash cache, then the effective cache size for the storage server is reduced. If the flash disk is used for flash log, then flash log is disabled on the disk thus reducing the effective flash log size. If the flash disk is used for grid disks, then the Oracle Automatic Storage Management (Oracle ASM) disks associated with these grid disks are automatically dropped with the FORCE
option from the Oracle ASM disk group, and a rebalance operation starts to restore the data redundancy.
The following procedure describes how to replace an FDOM due to disk failure on High Capacity storage servers that do not support online flash replacement. Replacing an NVMe drive on Extreme Flash storage servers is the same as replacing a physical disk: you can just remove the NVMe drive from the front panel and insert a new one. You do not need to shut down the storage server.
The new flash disk is automatically used by the system. If the flash disk is used for flash cache, then the effective cache size increases. If the flash disk is used for grid disks, then the grid disks are re-created on the new flash disk. If those grid disks were part of an Oracle ASM disk group, then they are added back to the disk group, and the data is rebalanced on them based on the disk group redundancy and ASM_POWER_LIMIT
parameter.
See Also:
- Performing a Hot Pluggable Replacement of a Flash Disk
- V$ASM_OPERATION in Oracle Database Reference
- ASM_POWER_LIMIT in Oracle Automatic Storage Management Administrator's Guide
- The appropriate PCIe Card User's Guide for your system, which is listed in "Related Documentation"
3.4.2 About Flash Disk Degraded Performance Statuses
If a flash disk has degraded performance, you might need to replace the disk.
You may need to replace a flash disk because the disk is has one of the following statuses:
warning - predictive failure
warning - poor performance
warning - write-through caching
warning - peer failure
Note:
For Oracle Exadata System Software releases earlier than release 11.2.3.2.2, the status isnot present
.
An alert is generated when a flash disk is in predictive failure
, poor performance
, write-through caching
or peer failure
status. The alert includes specific instructions for replacing the flash disk. If you have configured the system for alert notifications, then the alerts are sent by e-mail message to the designated address.
predictive failure
Flash disk predictive failure
status indicates that the flash disk will fail soon, and should be replaced at the earliest opportunity. If the flash disk is used for flash cache, then it continues to be used as flash cache. If the flash disk is used for grid disks, then the Oracle ASM disks associated with these grid disks are automatically dropped, and Oracle ASM rebalance relocates the data from the predictively failed disk to other disks.
When a flash disk has predictive failure
due to one flash disk, then the data is copied. If the flash disk is used for grid disks, then Oracle ASM re-partners the associated partner, and does a rebalance. If the flash disk is used for write back flash cache, then the data is flushed from the flash disks to the grid disks.
To identify a predictive failure
flash disk, use the following command:
CellCLI> LIST PHYSICALDISK WHERE DISKTYPE=flashdisk AND STATUS= \
'warning - predictive failure' DETAIL
name: FLASH_1_1
deviceName: /dev/nvme3n1
diskType: FlashDisk
luns: 1_1
makeModel: "Oracle Flash Accelerator F160 PCIe Card"
physicalFirmware: 8DV1RA13
physicalInsertTime: 2016-11-30T21:24:45-08:00
physicalSerial: CVMD519000251P6KGN
physicalSize: 1.4554837569594383T
slotNumber: "PCI Slot: 1; FDOM: 1"
status: warning - predictive failure
poor performance
Flash disk poor performance
status indicates that the flash disk demonstrates extremely poor performance, and should be replaced at the earliest opportunity. Starting with Oracle Exadata System Software release 11.2.3.2, an under-performing disk is automatically identified and removed from active configuration. If the flash disk is used for flash cache, then flash cache is dropped from this disk thus reducing the effective flash cache size for the storage server. If the flash disk is used for grid disks, then the Oracle ASM disks associated with the grid disks on this flash disk are automatically dropped with FORCE
option, if possible. If DROP...FORCE
cannot succeed due to offline partners, then the grid disks are automatically dropped normally, and Oracle ASM rebalance relocates the data from the poor performance disk to other disks.
Oracle Exadata Database Machine then runs a set of performance tests. When poor disk performance is detected by CELLSRV, the cell disk status changes to normal - confinedOnline
, and the physical disk status changes to warning - confinedOnline
. The following conditions trigger disk confinement:
- Disk stopped responding. The cause code in the storage alert log is CD_PERF_HANG.
- Slow cell disk such as the following:
- High service time threshold (cause code CD_PERF_SLOW_ABS)
- High relative service time threshold (cause code CD_PERF_SLOW_RLTV)
- High read or write latency such as the following:
- High latency on writes (cause code CD_PERF_SLOW_LAT_WT)
- High latency on reads (cause code CD_PERF_SLOW_LAT_RD)
- High latency on reads and writes (cause code CD_PERF_SLOW_LAT_RW)
- Very high absolute latency on individual I/Os happening frequently (cause code CD_PERF_SLOW_LAT_ERR)
- Errors such as I/O errors (cause code CD_PERF_IOERR).
If the disk problem is temporary and passes the tests, then it is brought back into the configuration. If the disk does not pass the tests, then it is marked as poor performance
, and Oracle Auto Service Request (ASR) submits a service request to replace the disk. If possible, Oracle ASM takes the grid disks offline for testing. If Oracle ASM cannot take the disks offline, then the cell disk status stays at normal - confinedOnline
until the disks can be taken offline safely.
To identify a poor performance
flash disk, use the following command:
CellCLI> LIST PHYSICALDISK WHERE DISKTYPE=flashdisk AND STATUS= \
'warning - poor performance' DETAIL
name: FLASH_1_4
diskType: FlashDisk
luns: 1_4
makeModel: "Sun Flash Accelerator F20 PCIe Card"
physicalFirmware: D20Y
physicalInsertTime: 2012-09-27T13:11:16-07:00
physicalSerial: 508002000092e70FMOD2
physicalSize: 22.8880615234375G
slotNumber: "PCI Slot: 1; FDOM: 3"
status: warning - poor performance
The disk status change is associated with the following entry in the cell alert history:
MESSAGE ID date_time info "Hard disk entered confinement status. The LUN
n_m changed status to warning - confinedOnline. CellDisk changed status to normal
- confinedOnline. Status: WARNING - CONFINEDONLINE Manufacturer: name Model
Number: model Size: size Serial Number: serial_number Firmware: fw_release
Slot Number: m Cell Disk: cell_disk_name Grid Disk: grid disk 1, grid disk 2
... Reason for confinement: threshold for service time exceeded"
The following would be logged in the storage server alert log:
CDHS: Mark cd health state change cell_disk_name with newState HEALTH_BAD_
ONLINE pending HEALTH_BAD_ONLINE ongoing INVALID cur HEALTH_GOOD
Celldisk entering CONFINE ACTIVE state with cause CD_PERF_SLOW_ABS activeForced: 0
inactiveForced: 0 trigger HistoryFail: 0, forceTestOutcome: 0 testFail: 0
global conf related state: numHDsConf: 1 numFDsConf: 0 numHDsHung: 0 numFDsHung: 0
...
Note:
In Oracle Exadata System Software releases earlier than release 11.2.3.2, use the CALIBRATE
command to identify a bad flash disk, and look for very low throughput and IOPS for each flash disk.
If a flash disk exhibits extremely poor performance, then it is marked as poor performance
. The flash cache on that flash disk is automatically disabled, and the grid disks on that flash disk are automatically dropped from the Oracle ASM disk group.
write-through caching
Flash disk write-through caching
status indicates the capacitors used to support data cache on the PCIe card have failed, and the card should be replaced as soon as possible.
peer failure
Flash disk peer failure
status indicates one of the flash disks on the same Sun Flash Accelerator PCIe card has failed or has a problem. For example, if FLASH_5_3 fails, then FLASH_5_0, FLASH_5_1, and FLASH_5_2 have peer failure status. The following is an example:
CellCLI> LIST PHYSICALDISK
36:0 L45F3A normal
36:1 L45WAE normal
36:2 L45WQW normal
...
FLASH_5_0 5L0034XM warning - peer failure
FLASH_5_1 5L0034JE warning - peer failure
FLASH_5_2 5L002WJH warning - peer failure
FLASH_5_3 5L002X4P failed
When CellSRV detects a predictive or peer failure in any flash disk used for write back flash cache and only one FDOM is bad, then the data on the bad FDOM is resilvered, and the data on the other three FDOMs is flushed. CellSRV then initiates an Oracle ASM rebalance for the disks if there are valid grid disks. The bad disk cannot be replaced until the tasks are completed. MS sends an alert when the disk can be replaced.
3.4.3 Replacing a Flash Disk Due to Flash Disk Problems
Oracle Exadata Storage Server is equipped with four PCIe cards. Each card has four flash disks (FDOMs) for a total of 16 flash disks. The four PCIe cards are present on PCI slot numbers 1, 2, 4, and 5. Starting with Oracle Exadata Database Machine X7, you can replace the PCIe cards without powering down the storage server. See Performing a Hot Pluggable Replacement of a Flash Disk.
In Oracle Exadata Database Machine X6 and earlier systems, the PCIe cards are not hot-pluggable. The Oracle Exadata Storage Server must be powered down before replacing the flash disks or cards.
Starting with Oracle Exadata Database Machine X7, each flash card on both High Capacity and Extreme Flash storage servers is a field-replaceable unit (FRU). The flash cards are also hot-pluggable, so you do not have to shut down the storage server before removing the flash card.
On Oracle Exadata Database Machine X5 and X6 systems, each flash card on High Capacity and each flash drive on Extreme Flash are FRUs. This means that there is no peer failure for these systems.
On Oracle Exadata Database Machine X3 and X4 systems, because the flash card itself is a FRU, if any FDOMs were to fail, the Oracle Exadata System Software would automatically put the rest of FDOMs on that card to peer failure so that the data can be moved out to prepare for the flash card replacement.
On Oracle Exadata Database Machine V2 and X2 systems, each FDOM is a FRU. There is no peer failure for flash for these systems.
Determining when to proceed with disk replacement depends on the release, as described in the following:
-
For Oracle Exadata System Software releases earlier than 11.2.3.2:
Wait until the Oracle ASM disks have been successfully dropped by querying the
V$ASM_DISK_STAT
view before proceeding with the flash disk replacement. If the normal drop did not complete before the flash disk fails, then the Oracle ASM disks are automatically dropped with theFORCE
option from the Oracle ASM disk group. If theDROP
command did not complete before the flash disk fails, then refer to Replacing a Flash Disk Due to Flash Disk Failure. -
For Oracle Exadata System Software releases 11.2.3.2 and later:
An alert is sent when the Oracle ASM disks have been dropped, and the flash disk can be safely replaced. If the flash disk is used for write-back flash cache, then wait until none of the grid disks are cached by the flash disk. Use the following command to check the
cachedBy
attribute of all the grid disks. The cell disk on the flash disk should not appear in any grid disk'scachedBy
attribute.CellCLI> LIST GRIDDISK ATTRIBUTES name, cachedBy
If the flash disk is used for both grid disks and flash cache, then wait until receiving the alert, and the cell disk is not shown in any grid disk's
cachedBy
attribute.
The following procedure describes how to replace a flash disk on High Capacity storage servers for Oracle Exadata Database Machine X6 and earlier due to disk problems.
Note:
On Extreme Flash storage servers for Oracle Exadata Database Machine X6 and all storage servers for Oracle Exadata Database Machine X7 and later, you can just remove the flash disk from the front panel and insert a new one. You do not need to shut down the storage server.The new flash disk is automatically used by the system. If the flash disk is used for flash cache, then the effective cache size increases. If the flash disk is used for grid disks, then the grid disks are re-created on the new flash disk. If those gird disks were part of an Oracle ASM disk group, then they are added back to the disk group, and the data is rebalanced on them based on the disk group redundancy and the ASM_POWER_LIMIT
parameter.
Related Topics
See Also:
- Oracle Automatic Storage
Management Administrator's Guide for information about the
ASM_POWER_LIMIT
parameter - The appropriate PCIe Card User's Guide for your system, which is listed in Related Documentation
3.4.4 Performing a Hot Pluggable Replacement of a Flash Disk
Starting with Oracle Exadata Database Machine X7, flash disks are hot-pluggable on both Extreme Flash (EF) and High Capacity (HC) storage servers.
Also, for Oracle Exadata Database Machine X6 and earlier, the flash devices on EF storage servers are hot-pluggable. However, for HC storage servers on Oracle Exadata Database Machine X6 and earlier systems, you must power down the storage servers before replacing the flash disks.
To replace a hot-pluggable flash disk device:
Related Topics
3.4.5 Enabling Write Back Flash Cache
Write operations serviced by flash instead of by disk are referred to as write-back flash cache.
Starting with Oracle Exadata System Software release 11.2.3.2.1, Exadata Smart Flash Cache can transparently cache frequently-accessed data to fast solid-state storage, improving query response times and throughput.
- Enable Write Back Flash Cache for 11.2.3.3.1 or Higher
Enable write back Flash Cache on the storage servers to improve query response times and throughput. - Enabling Write Back Flash Cache on a Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on a rolling basis. - Enabling Write Back Flash Cache in a Non-Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on a non-rolling basis.
3.4.5.1 Enable Write Back Flash Cache for 11.2.3.3.1 or Higher
Enable write back Flash Cache on the storage servers to improve query response times and throughput.
For Oracle Exadata System Software release 11.2.3.3.1 or higher, you do not have to stop cell services or inactivate grid disks when changing the Flash Cache from Write Through mode to Write Back mode.
Note:
Any time the Flash Cache is dropped and re-created, there is a performance impact for database operations. While the Flash Cache is being repopulated, there are more cache misses, which impacts database performance.Parent topic: Enabling Write Back Flash Cache
3.4.5.2 Enabling Write Back Flash Cache on a Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on a rolling basis.
To modify the Flash Cache attribute from writethrough
to writeback
, the Flash Cache must be dropped first. For Oracle Exadata System Software releases prior to 11.2.3.3.1, you must stop cell services or inactivate grid disks when enabling Write Back Flash Cache.
There is a shell script to automate enabling and disabling Write Back Flash Cache. Refer to My Oracle Support note 1500257.1 for the script and additional information.
Note:
Any time the Flash Cache is dropped and re-created, there is a performance impact for database operations. While the Flash Cache is being repopulated, there are more cache misses, which impacts database performance.Oracle Grid Infrastructure homes and Oracle Database homes must at release 11.2.0.3 BP9 or higher to use write-back Flash Cache. Refer to My Oracle Support note 888828.1 for the minimum release requirements for Oracle Exadata System Software, Oracle Grid Infrastructure home, and Oracle Database home.
3.4.5.3 Enabling Write Back Flash Cache in a Non-Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on a non-rolling basis.
The cell services must be shut down before changing the flashCacheMode
attribute. For Oracle Exadata System Software releases prior to 11.2.3.3.1, you must stop cell services when disabling Write Back Flash Cache in a non-rolling manner.
There is a shell script to automate enabling and disabling write back Flash Cache. Refer to My Oracle Support note 1500257.1 for the script and additional information.
Oracle Grid Infrastructure homes and Oracle Database homes must at release 11.2.0.3 BP9 or higher to use write-back Flash Cache. Refer to My Oracle Support note 888828.1 for the minimum release requirements for Oracle Exadata System Software, Oracle Grid Infrastructure home, and Oracle Database home.
Parent topic: Enabling Write Back Flash Cache
3.4.6 Disabling Write Back Flash Cache
You can disable the Write-Back Flash Cache by enabling Write-Through Flash Cache.
Starting with Oracle Exadata System Software release 11.2.3.2.1, Exadata Smart Flash Cache can transparently cache frequently-accessed data to fast solid-state storage, improving query response times and throughput.
Write operations serviced by flash instead of by disk are referred to as write back flash cache.
- Disable Write-Back Flash Cache Along With Write-Back PMEM Cache
- Disable Write Back Flash Cache for 11.2.3.3.1 or Higher
You can disable write back Flash Cache on the storage servers by changing the mode to Write Through. - Disabling Write Back Flash Cache on a Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on each storage server using a rolling method. - Disabling Write Back Flash Cache on a Non-Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on a non-rolling basis.
3.4.6.1 Disable Write-Back Flash Cache Along With Write-Back PMEM Cache
Before Oracle Exadata System Software release 23.1.0, write-back PMEM cache is only supported in conjunction with write-back flash cache. Consequently, if write-back PMEM Cache is enabled, you must also disable write-back PMEM Cache to disable write-back flash cache.
This requirement only applies prior to Oracle Exadata System Software release 23.1.0 because PMEM cache only operates in write-through mode starting with Oracle Exadata System Software release 23.1.0.
Note:
To reduce the performance impact on your applications, change the cache mode during a period of reduced workload.
The following command examples use a text file named cell_group
that contains
the host names of the storage servers that are the subject of the procedure.
Parent topic: Disabling Write Back Flash Cache
3.4.6.2 Disable Write Back Flash Cache for 11.2.3.3.1 or Higher
You can disable write back Flash Cache on the storage servers by changing the mode to Write Through.
With release 11.2.3.3.1 or higher, you do not have to stop the CELLSRV process or inactivate grid disks.
Note:
To reduce the performance impact on the application, disable the write back flash cache during a period of reduced workload.Parent topic: Disabling Write Back Flash Cache
3.4.6.3 Disabling Write Back Flash Cache on a Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on each storage server using a rolling method.
The cell services must be shut down before changing the flashCacheMode
attribute. The cell services can shut down on rolling basis. The flash cache must be flushed and dropped before changing the attribute to writethrough
. After the flush operation begins, all caching to the flash cache stops. Ensure cell services resynchronization is complete on the current storage server before changing the next storage server.
There is a shell script to automate enabling and disabling write back Flash Cache. Refer to My Oracle Support note 1500257.1 for the script and additional information.
Parent topic: Disabling Write Back Flash Cache
3.4.6.4 Disabling Write Back Flash Cache on a Non-Rolling Basis for Software Versions Lower Than 11.2.3.3.1
You can enable Write Back Flash Cache on a non-rolling basis.
When changing the Flash Cache mode on a non-rolling basis, ensure the entire cluster is shut down, including the Oracle Clusterware stack and all the databases. The cell services must be shut down before changing the flashCacheMode
attribute. The Flash Cache must be flushed and dropped before changing the attribute to writethrough
. The Flash Cache flush operation can be performed prior to shutting down the entire cluster. After the flush operation begins, all caching to the Flash Cache stops.
Parent topic: Disabling Write Back Flash Cache
3.4.7 Enabling Flash Cache Compression
Flash cache compression can be enabled on Oracle Exadata Database Machine X4-2, Oracle Exadata Database Machine X3-2, and Oracle Exadata Database Machine X3-8 Full Rack systems. Oracle Exadata Database Machine X5-2, X5-8, and later systems do not have flash cache compression. Flash cache compression dynamically increases the logical capacity of the flash cache by transparently compressing user data as it is loaded into the flash cache.
Note:
-
Oracle Advanced Compression Option is required to enable flash cache compression.
-
User data is not retained when enabling flash cache compression.
The following procedure describes how to enable flash cache compression:
3.4.8 Monitoring Exadata Smart Flash Cache Usage Statistics
Use the following methods to monitor Exadata Smart Flash Cache usage:
Related Topics
3.4.9 Disabling Flash Cache Compression
Flash cache compression can be disabled on Oracle Exadata Database Machine X4-2, Oracle Exadata Database Machine X3-2, and Oracle Exadata Database Machine X3-8 Full Rack systems. Oracle Exadata Database Machine X5-2, X5-8, and later systems do not have flash cache compression.
Note:
-
User data is not retained when disabling flash cache compression.
The following procedure describes how to disable flash cache compression:
3.5 Maintaining PMEM Devices on Oracle Exadata Storage Servers
Persistent memory (PMEM) devices reside in Exadata X8M-2 and X9M-2 storage server models with High Capacity (HC) or Extreme Flash (EF) storage.
If a PMEM device fails, Oracle Exadata System Software isolates the failed device and automatically recovers the cache using the device.
If the cache is in write-back mode, the recovery operation, also known as
resilvering, restores the lost data by reading a mirrored copy. During resilvering, the
grid disk status is ACTIVE -- RESILVERING WORKING
. If the cache is in
write-through mode, then the data in the failed PMEM device is already stored in the
data grid disk, and no recovery is required.
- Replacing a PMEM Device Due to Device Failure
If the PMEM device has a status ofFailed
, you should replace the PMEM device on the Oracle Exadata Storage Server. - Replacing a PMEM Device Due to Degraded Performance
If a PMEM device has degraded performance, you might need to replace the module. - Enabling and Disabling Write-Back PMEM Cache
Parent topic: Maintaining Oracle Exadata Storage Servers
3.5.1 Replacing a PMEM Device Due to Device Failure
If the PMEM device has a status of Failed
, you should replace the PMEM device on the Oracle Exadata Storage Server.
A PMEM fault could cause server to reboot. The failed device should be
replaced with a new PMEM device at the earliest opportunity. Until the PMEM device
is replaced, the corresponding cache size is reduced. If the PMEM device is used for
commit acceleration (XRMEMLOG
or PMEMLOG
), then
the size of the corresponding commit accelerator is also reduced.
An alert is generated when a PMEM device failure is detected. The alert message includes the slot number and cell disk name. If you have configured the system for alert notification, then an alert is sent by e-mail message to the designated address.
To identify a failed PMEM device, you can also use the following command:
CellCLI> LIST PHYSICALDISK WHERE disktype=PMEM AND status=failed DETAIL
name: PMEM_0_1
diskType: PMEM
luns: P0_D1
makeModel: "Intel NMA1XBD128GQS"
physicalFirmware: 1.02.00.5365
physicalInsertTime: 2019-09-28T11:29:13-07:00
physicalSerial: 8089-A2-1838-00001234
physicalSize: 126.375G
slotNumber: "CPU: 0; DIMM: 1"
status: failed
In the above output, the slotNumber
shows the socket
number and DIMM slot number.
The new PMEM device is automatically used by the system. If the PMEM device is used for caching, then the effective cache size increases. If the PMEM device is used for commit acceleration, then commit acceleration is enabled on the device.
3.5.2 Replacing a PMEM Device Due to Degraded Performance
If a PMEM device has degraded performance, you might need to replace the module.
If degraded performance is detected on a PMEM device, the module status is set to warning - predictive
failure
and an alert is generated. The alert includes specific instructions
for replacing the PMEM device. If you have configured the
system for alert notifications, then the alerts are sent by e-mail message to the
designated address.
The predictive failure
status indicates that the PMEM device will fail soon, and should be replaced at the
earliest opportunity. No new data is cached in the PMEM device until it is replaced.
To identify a PMEM device with the status
predictive failure
, you can also use the following command:
CellCLI> LIST PHYSICALDISK WHERE disktype=PMEM AND status='warning - predictive failure' DETAIL
name: PMEM_0_6
diskType: PMEM
luns: P0_D6
makeModel: "Intel NMA1XBD128GQS"
physicalFirmware: 1.02.00.5365
physicalInsertTime: 2019-11-30T21:24:45-08:00
physicalSerial: 8089-A2-1838-00001234
physicalSize: 126.375G
slotNumber: "CPU: 0; DIMM: 6"
status: warning - predictive failure
You can also locate the PMEM device using the information
in the LIST DISKMAP
command:
CellCLI> LIST DISKMAP
Name PhysicalSerial SlotNumber Status PhysicalSize
CellDisk DevicePartition GridDisks
PMEM_0_1 8089-a2-0000-00000460 "CPU: 0; DIMM: 1" normal 126G
PM_00_cel01 /dev/dax5.0 PMEMCACHE_PM_00_cel01
PMEM_0_3 8089-a2-0000-000004c2 "CPU: 0; DIMM: 3" normal 126G
PM_02_cel01 /dev/dax4.0 PMEMCACHE_PM_02_cel01
PMEM_0_5 8089-a2-0000-00000a77 "CPU: 0; DIMM: 5" normal 126G
PM_03_cel01 /dev/dax3.0 PMEMCACHE_PM_03_cel01
PMEM_0_6 8089-a2-0000-000006ff "CPU: 0; DIMM: 6" warning - 126G
PM_04_cel01 /dev/dax0.0 PMEMCACHE_PM_04_cel01
PMEM_0_8 8089-a2-0000-00000750 "CPU: 0; DIMM: 8" normal 126G
PM_05_cel01 /dev/dax1.0 PMEMCACHE_PM_05_cel01
PMEM_0_10 8089-a2-0000-00000103 "CPU: 0; DIMM: 10" normal 126G
PM_01_cel01 /dev/dax2.0 PMEMCACHE_PM_01_cel01
PMEM_1_1 8089-a2-0000-000008f6 "CPU: 1; DIMM: 1" normal 126G
PM_06_cel01 /dev/dax11.0 PMEMCACHE_PM_06_cel01
PMEM_1_3 8089-a2-0000-000003bb "CPU: 1; DIMM: 3" normal 126G
PM_08_cel01 /dev/dax10.0 PMEMCACHE_PM_08_cel01
PMEM_1_5 8089-a2-0000-00000708 "CPU: 1; DIMM: 5" normal 126G
PM_09_cel01 /dev/dax9.0 PMEMCACHE_PM_09_cel01
PMEM_1_6 8089-a2-0000-00000811 "CPU: 1; DIMM: 6" normal 126G
PM_10_cel01 /dev/dax6.0 PMEMCACHE_PM_10_cel01
PMEM_1_8 8089-a2-0000-00000829 "CPU: 1; DIMM: 8" normal 126G
PM_11_cel01 /dev/dax7.0 PMEMCACHE_PM_11_cel01
PMEM_1_10 8089-a2-0000-00000435 "CPU: 1; DIMM: 10" normal 126G
PM_07_cel01 /dev/dax8.0 PMEMCACHE_PM_07_cel01
If the PMEM device is used for write-back
caching, then the data is flushed from the PMEM device to
the flash cache. To ensure that data is flushed from the PMEM device, check the cachedBy
attribute of all the grid disks and
ensure that the affected PMEM device is not listed.
The new PMEM device is automatically used by the system. If the PMEM device is used for caching, then the effective cache size increases. If the PMEM device is used for commit acceleration, then commit acceleration is enabled on the device.
3.5.3 Enabling and Disabling Write-Back PMEM Cache
Prior to Oracle Exadata System Software release 23.1.0, you can configure PMEM cache to operate in write-back mode. Also known as write-back PMEM cache, this mode enables the cache to service write operations.
Note:
The best practice recommendation is to configure PMEM cache in write-through mode. This configuration provides the best performance and availability.
Commencing with Oracle Exadata System Software release 23.1.0, PMEM cache only operates in write-through mode.
- Enable Write-Back PMEM Cache
- Disable Write-Back PMEM Cache
Use these steps if you need to disable Write-Back PMEM cache on the storage servers.
3.5.3.1 Enable Write-Back PMEM Cache
Write-back PMEM cache is only supported in conjunction with write-back flash cache. Consequently, to enable write-back PMEM cache you must also enable write-back flash cache.
Note:
Commencing with Oracle Exadata System Software release 23.1.0, you cannot enable write-back PMEM cache because PMEM cache only operates in write-through mode.
Note:
To reduce the performance impact on your applications, change the cache mode during a period of reduced workload.
The following command examples use a text file named cell_group
that contains
the host names of the storage servers that are the subject of the procedure.
Parent topic: Enabling and Disabling Write-Back PMEM Cache
3.5.3.2 Disable Write-Back PMEM Cache
Use these steps if you need to disable Write-Back PMEM cache on the storage servers.
You do not have to stop the cellsrv process or inactivate grid disks when disabling Write-Back PMEM cache. However, to reduce the performance impact on the application, disable the Write-Back PMEM cache during a period of reduced workload.
Parent topic: Enabling and Disabling Write-Back PMEM Cache
3.6 Maintaining the M.2 Disks of Oracle Exadata Storage Server
Oracle Exadata X7 and later systems come with two internal M.2 devices that contain the system area.
In all previous systems, the first two disks of the Oracle Exadata Storage Server are system disks and the portions on these system disks are referred to as the system area.
Note:
Oracle Exadata Rack and Oracle Exadata Storage Servers can remain online and available while replacing an M.2 disk.- Monitoring the Status of M.2 Disks
You can monitor the status of a M.2 disk by checking its attributes with the CellCLILIST PHYSICALDISK
command. - Replacing a M.2 Disk Due to Failure or Other Problems
Parent topic: Maintaining Oracle Exadata Storage Servers
3.6.1 Monitoring the Status of M.2 Disks
You can monitor the status of a M.2 disk by checking its attributes with the CellCLI LIST PHYSICALDISK
command.
The disk firmware maintains the error counters, and marks a drive with Predictive Failure when the disk is about to fail. The drive, not the cell software, determines if it needs replacement.
3.6.2 Replacing a M.2 Disk Due to Failure or Other Problems
Failure of a M.2 disks reduces redundancy of the system area, and can impact patching, imaging, and system rescue. Therefore, the disk should be replaced with a new disk as soon as possible. When a M.2 disk fails, the storage server automatically and transparently switches to using the software stored on the inactive system disk, making it the active system disk.
An Exadata alert is generated when an M.2 disk fails. The alert includes specific instructions for replacing the disk. If you have configured the system for alert notifications, then the alert is sent by e-mail to the designated address.
M.2 disk is hot-pluggable and can be replaced when the power is on.
After the M.2 disk is replaced, Oracle Exadata System Software automatically adds the new device to the system partition and starts the rebuilding process.
3.7 Resizing Grid Disks
You can resize grid disks and Oracle ASM disk groups to shrink one with excess free space and increase the size of another that is near capacity.
-
For internal backups: allocation of available space is 40% for the DATA disk groups, and 60% for the RECO disk groups.
-
For external backups: allocation of available space is 80% for the DATA disk group, and 20% for the RECO disk group.
If your system has no free space available on the cell disks and one disk group, for example RECO, has plenty of free space, then you can resize the RECO disk group to a smaller size and reallocate the free space to the DATA disk group. The free space available after shrinking the RECO disk group is at a non-contiguous offset from the existing space allocations for the DATA disk group. Grid disks can use space anywhere on the cell disks and do not have to be contiguous.
If you are expanding the grid disks and the cell disks already have sufficient space to expand the existing grid disks, then you do not need to first resize an existing disk group. You would skip steps 2 and 3 below where the example shows the RECO disk group and grid disks are shrunk (you should still verify the cell disks have enough free space before growing the DATA grid disks). The amount of free space the administrator should reserve depends on the level of failure coverage.
If you are shrinking the size of the grid disks, you should understand how space is reserved for mirroring. Data is protected by Oracle ASM using normal or high redundancy to create one or two copies of data, which are stored as file extents. These copies are stored in separate failure groups. A failure in one failure group does not affect the mirror copies, so data is still accessible.
When a failure occurs, Oracle ASM re-mirrors, or rebalances, any extents that are not accessible so that redundancy is reestablished. For the re-mirroring process to succeed, sufficient free space must exist in the disk group to allow creation of the new file extent mirror copies. If there is not enough free space, then some extents will not be re-mirrored and the subsequent failure of the other data copies will require the disk group to be restored from backup. Oracle ASM sends an error when a re-mirror process fails due to lack of space.
You must be using Oracle Exadata System Software release 12.1.2.1.0 or higher, or have the patch for bug 19695225 applied to your software.
This procedure for resizing grid disks applies to bare metal and virtual machine (VM) deployments.
- Determine the Amount of Available Space
To increase the size of the disks in a disk group you must either have unallocated disk space available, or you have to reallocate space currently used by a different disk group. - Shrink the Oracle ASM Disks in the Donor Disk Group
If there is no free space available on the cell disks, you can reduce the space used by one disk group to provide additional disk space for a different disk group. - Shrink the Grid Disks in the Donor Disk Group
After shrinking the disks in the Oracle ASM disk group, you then shrink the size of the grid disks on each cell. - Increase the Size of the Grid Disks Using Available Space
You can increase the size used by the grid disks if there is unallocated disk space either already available, or made available by shrinking the space used by a different Oracle ASM disk group. - Increase the Size of the Oracle ASM Disks
You can increase the size used by the Oracle ASM disks after increasing the space allocated to the associated grid disks.
Related Topics
- Understanding ASM Capacity and Reservation of Free Space in Exadata (My Oracle Support Doc ID 1551288.1)
- Bug 19695225 - Running Many Create or Alter Griddisk Commands Over Time Causes Cell Disk Metadata Corruption (ORA-600 [addNewSegmentsToGDisk_2]) and Loss of Cell Disk Content (My Oracle Support Doc ID 1991445.1)
Parent topic: Maintaining Oracle Exadata Storage Servers
3.7.1 Determine the Amount of Available Space
To increase the size of the disks in a disk group you must either have unallocated disk space available, or you have to reallocate space currently used by a different disk group.
You can also use a script available in "Script to Calculate New Grid Disk and Disk Group Sizes in Exadata (My Oracle Support Doc ID 1464809.1)" to assist in determining how much free space is available to shrink a disk group.
Parent topic: Resizing Grid Disks
3.7.2 Shrink the Oracle ASM Disks in the Donor Disk Group
If there is no free space available on the cell disks, you can reduce the space used by one disk group to provide additional disk space for a different disk group.
Parent topic: Resizing Grid Disks
3.7.3 Shrink the Grid Disks in the Donor Disk Group
After shrinking the disks in the Oracle ASM disk group, you then shrink the size of the grid disks on each cell.
Parent topic: Resizing Grid Disks
3.7.4 Increase the Size of the Grid Disks Using Available Space
You can increase the size used by the grid disks if there is unallocated disk space either already available, or made available by shrinking the space used by a different Oracle ASM disk group.
This task is a continuation of an example where space in the RECOC1 disk group is being reallocated to the DATAC1 disk group. If you already have sufficient space to expand an existing disk group, then you do not need to reallocate space from a different disk group.
Instead of increasing the size of the DATA disk group, you could instead create new disk groups with the new free space or keep it free for future use. In general, Oracle recommends using the smallest number of disk groups needed (typically DATA, RECO, and DBFS_DG) to give the greatest flexibility and ease of administration. However, there may be cases, perhaps when using virtual machines or consolidating many databases, where additional disk groups or available free space for future use may be desired.
If you decide to leave free space on the grid disks in reserve for future use, please see the My Oracle Support Note 1684112.1 for the steps on how to allocate free space to an existing disk group at a later time.
3.7.5 Increase the Size of the Oracle ASM Disks
You can increase the size used by the Oracle ASM disks after increasing the space allocated to the associated grid disks.
Related Topics
Parent topic: Resizing Grid Disks
3.8 Using the Oracle Exadata System Software Rescue Procedure
In the rare event that the system disks fail simultaneously, you must use the Oracle Exadata Storage Server rescue procedure to recover the system.
- About the Oracle Exadata System Software Rescue Procedure
The rescue procedure is necessary when system disks fail, the operating system has a corrupt file system, or there was damage to the boot area. - Performing the Rescue Procedure
You can use the rescue procedure to recover the Oracle Exadata Storage Server system software. - Configuring Oracle Exadata Storage Servers After Rescue
After a successful rescue, you must configure the cell. If the data partitions were preserved, then the cell disks were imported automatically during the rescue procedure. - Configuring Oracle Exadata Database Machine Eighth Rack Storage Servers After Rescue
For storage servers that are part of an Eighth Rack system, after a successful rescue, you must configure the cell using these steps. - Re-creating a Damaged CELLBOOT USB Flash Drive
Parent topic: Maintaining Oracle Exadata Storage Servers
3.8.1 About the Oracle Exadata System Software Rescue Procedure
The rescue procedure is necessary when system disks fail, the operating system has a corrupt file system, or there was damage to the boot area.
If only one system disk fails, then use CellCLI commands to recover.
If you are using normal redundancy, then there is only one mirror copy for the cell being rescued. The data may be irrecoverably lost if that single mirror also fails during the rescue procedure. Oracle recommends that you take a complete backup of the data on the mirror copy, and immediately take the mirror copy cell offline to prevent any new data changes to it prior to attempting a rescue. This ensures that all data residing on the grid disks on the failed cell and its mirror copy is inaccessible during rescue procedure.
The Oracle Automatic Storage Management (Oracle ASM) disk repair timer has a default repair time of 3.6 hours. If you know that you cannot perform the rescue procedure within that time frame, then you should use the Oracle ASM rebalance procedure to rebalance the disk until you can do the rescue procedure.
When using high redundancy disk groups, such as having more than one mirror copy in Oracle ASM for all the grid disks of the failed cell, then take the failed cell offline. Oracle ASM automatically drops the grid disks on the failed cell after the configured Oracle ASM time out, and starts rebalancing data using mirror copies. The default timeout is two hours. If the cell rescue takes more than two hours, then you must re-create the grid disks on the rescued cells in Oracle ASM.
Caution:
Use the rescue procedure with extreme caution. Incorrectly using the procedure can cause data loss.It is important to note the following when using the rescue procedure:
-
The rescue procedure can potentially rewrite some or all of the disks in the cell. If this happens, then you can lose all the content on those disks without possibility of recovery.
Use extreme caution when using this procedure, and pay attention to the prompts. Ideally, you should use the rescue procedure only with assistance from Oracle Support Services, and when you have decided that you can afford the loss of data on some or all of the disks.
-
The rescue procedure does not destroy the contents of the data disks or the contents of the data partitions on the system disks unless you explicitly choose to do so during the rescue procedure.
-
Starting in Oracle Exadata System Software release 11.2, the rescue procedure restores the Oracle Exadata System Software to the same release. This includes any patches that existed on the cell as of the last successful boot. Note the following about using the rescue procedure:
-
Cell configuration information, such as alert configurations, SMTP information, administrator e-mail address, and so on is not restored.
-
The network configuration that existed at the end of last successful run of
/usr/local/bin/ipconf
utility is restored. -
The SSH identities for the cell, and the
root
,celladmin
andcellmonitor
users are restored. -
Integrated Lights Out Manager (ILOM) configurations for Oracle Exadata Storage Servers are not restored. Typically, ILOM configurations remain undamaged even in case of Oracle Exadata System Software failures.
-
-
The rescue procedure does not examine or reconstruct data disks or data partitions on the system disks. If there is data corruption on the grid disks, then do not use the rescue procedure. Instead use the rescue procedure for Oracle Database and Oracle ASM.
After a successful rescue, you must reconfigure the cell, and if you had chosen to preserve the data, then import the cell disks. If you chose not to preserve the data, then you should create new cell disks, and grid disks.
3.8.2 Performing the Rescue Procedure
You can use the rescue procedure to recover the Oracle Exadata Storage Server system software.
3.8.3 Configuring Oracle Exadata Storage Servers After Rescue
After a successful rescue, you must configure the cell. If the data partitions were preserved, then the cell disks were imported automatically during the rescue procedure.
Related Topics
3.8.4 Configuring Oracle Exadata Database Machine Eighth Rack Storage Servers After Rescue
For storage servers that are part of an Eighth Rack system, after a successful rescue, you must configure the cell using these steps.
In Oracle Exadata System Software release 11.2.3.3 and later, no extra steps are needed after cell rescue.
3.8.5 Re-creating a Damaged CELLBOOT USB Flash Drive
If the CELLBOOT USB flash drive is lost or damaged, then you can create a new one using the following procedure:
Note:
To create a USB flash drive for a machine running Oracle Exadata Storage Server Software release 12.1.2.1.0 or later requires a machine running Oracle Linux 6.
3.9 Changing Existing Elastic Configurations for Storage Cells
You can modify the capacity of your Oracle Exadata using elastic configuration.
- Adding a Cell Node
In this scenario, you want to add a new storage server (or cell) to an existing Oracle Exadata that includes disk groups. - Adding a New Storage Server to an Eighth Rack Cluster
Perform the following steps to add a new Oracle Exadata X7 or later storage server to an existing Oracle Exadata X7 or later Eighth Rack. - Adding Storage Cells using OEDACLI
OEDACLI provides the interface to perform elastic storage expansion for different configurations such as Bare Metal, single Oracle VM or multiple Oracle VMs. - Expanding an Existing Exadata Storage Grid
- Dropping a Storage Server from an Existing Disk Group or Storage Grid
You can remove a storage server from an existing Oracle Exadata Rack. - Dropping Storage Servers using OEDACLI
OEDACLI provides the interface to drop storage servers for different configuration such as Bare Metal, single Oracle VM or multiple Oracle VMs.
Parent topic: Maintaining Oracle Exadata Storage Servers
3.9.1 Adding a Cell Node
In this scenario, you want to add a new storage server (or cell) to an existing Oracle Exadata that includes disk groups.
3.9.2 Adding a New Storage Server to an Eighth Rack Cluster
Perform the following steps to add a new Oracle Exadata X7 or later storage server to an existing Oracle Exadata X7 or later Eighth Rack.
3.9.3 Adding Storage Cells using OEDACLI
OEDACLI provides the interface to perform elastic storage expansion for different configurations such as Bare Metal, single Oracle VM or multiple Oracle VMs.
Prerequisites
- All the new storage servers must be installed in and connected into the physical rack.
- All the new storage servers must have the management and RDMA Network Fabric networks configured.
- OEDACLI must run on a machine that has network access to the database servers (bare metal or guest domains) and the storage servers.
3.9.4 Expanding an Existing Exadata Storage Grid
In this scenario, you have an Exadata storage cell in an Exadata rack, and you want to add the storage cell to an Exadata storage grid that you most probably want to expand.
3.9.5 Dropping a Storage Server from an Existing Disk Group or Storage Grid
You can remove a storage server from an existing Oracle Exadata Rack.
3.9.6 Dropping Storage Servers using OEDACLI
OEDACLI provides the interface to drop storage servers for different configuration such as Bare Metal, single Oracle VM or multiple Oracle VMs.
The procedure to remove the grid disks from the disk groups and drop the objects on the storage server is implemented through only a few commands. There is only one rebalance operation triggered, regardless of the number of storage cells dropped.
Prerequisites
- A valid OEDA XML configuration file, reflecting an exact list of the compute nodes and storage cells used in the environment to be expanded. The first step in this task generates a new OEDA XML configuration file to ensure that the current configuration is used.
- OEDACLI must run on a machine that has network access to the database servers (bare metal or guest domains) and the storage servers.
3.10 Managing Disk Controller Batteries
This section applies only to Exadata systems prior to X6 that use batteries. Newer systems have CVPM02 (Cache Vault), which is a super cap and not a battery.
- About Disk Controller Batteries
The disk controllers in Oracle Exadata storage servers and database servers have battery-backed write cache to accelerate write performance. - Monitoring Batteries in the Database Servers
- Replacing Batteries in Disk Controllers
Parent topic: Maintaining Oracle Exadata Storage Servers
3.10.1 About Disk Controller Batteries
The disk controllers in Oracle Exadata storage servers and database servers have battery-backed write cache to accelerate write performance.
Note:
This applies only to Exadata systems prior to X6 that use batteries. Newer systems have CVPM02 (Cache Vault), which is a super cap and not a battery.If the battery charge capacity degrades such that the battery can no longer protect the cached data for a power loss of 48 hours or more, then the write cache is disabled and the disk controller switches to write through mode. This results in reduced write performance, but there is no data loss. Oracle Exadata Storage Servers generate an alert when battery charge capacity is insufficient or the temperature is high, and when the battery should be replaced.
Battery charge capacity degrades over time, and its life expectancy is inversely proportional to the operating temperature. The worst case life expectancy of the battery in Oracle Exadata Rack is as follows:
Inlet Ambient Temperature | Battery Lifetime |
---|---|
< 25 degrees Celsius (77 degrees Fahrenheit) |
3 years |
< 32 degrees Celsius (89.6 degrees Fahrenheit) |
2 years |
Parent topic: Managing Disk Controller Batteries
3.10.2 Monitoring Batteries in the Database Servers
Note:
Exadata Storage Servers generate an alert when battery charge capacity is insufficient or the temperature is high, and when the battery should be replaced.
The battery charge capacity and battery temperature in the database servers can be monitored using the following commands:
Note:
If you are running Oracle Exadata System Software 19.1.0 or later, substitute/opt/MegaRAID/storcli/storcli64
for /opt/MegaRAID/MegaCli/MegaCli64
in the following commands:
# /opt/MegaRAID/MegaCli/MegaCli64 -AdpBbuCmd -a0 | grep "Full Charge" -A5 | sort \
| grep Full -A1
The following is an example of the output from the command:Full Charge Capacity: 1357 mAh
Max Error: 2 %
Proactive battery replacement should be done on batteries that show capacity less than 800 mAh and have maximum error less than 10%. Immediately replace any battery that has less than 674 mAh or has maximum error more than 10%./opt/MegaRAID/MegaCli/MegaCli64 -AdpBbuCmd -a0 | grep BatteryType; \
/opt/MegaRAID/MegaCli/MegaCli64 -AdpBbuCmd -a0 | grep -i temper
The following is an example of the output from the command:BatteryType: iBBU08
Temperature: 38 C
Temperature : OK
Over Temperature : No
If the battery temperature is greater than or equal to 55 degrees Celsius, then determine the cause, and correct the problem.Parent topic: Managing Disk Controller Batteries
3.10.3 Replacing Batteries in Disk Controllers
If the battery charge capacity in the disk controllers falls below the minimum threshold, then Oracle will replace the failed batteries at no extra charge, if the system is covered either by the Oracle Premier Support for Systems or occurs during the warranty period.
For customers with Premier Support for Systems, Oracle attempts to proactively replace the batteries in Oracle Exadata Rack before the end of the estimated lifetime, on a best efforts basis.
Parent topic: Managing Disk Controller Batteries
3.11 Managing F20 PCIe Energy Storage Modules
Sun Flash Accelerator F20 PCIe cards are used in Oracle Exadata X3 models.
- About F20 PCIe Energy Storage Modules
The Sun Flash Accelerator F20 PCIe card includes an energy storage module (ESM) to ensure data integrity during a power interruption, functioning similar to a battery backup. - Replacing Flash ESM
Parent topic: Maintaining Oracle Exadata Storage Servers
3.11.1 About F20 PCIe Energy Storage Modules
The Sun Flash Accelerator F20 PCIe card includes an energy storage module (ESM) to ensure data integrity during a power interruption, functioning similar to a battery backup.
Sun Flash Accelerator F20 PCIe cards accelerate performance in Oracle Exadata Rack by caching frequently-accessed Oracle Database data and avoids the need to do physical I/O to the disk in Exadata Storage Server. Write operations to the flash cards are temporarily staged in volatile local DRAM memory on the card to speed write operations. The data in the DRAM is protected by an Energy Storage Module (ESM) which provides enough electrical power, in the event of a power failure, to move the data in the DRAM to the local flash.
The flash modules used in Oracle Exadata X3 systems have an expected endurance of 10 years or more, even in write intensive applications. Flash endurance is determined primarily by the total data written to flash across many years, as well as the type of data written. No application runs at maximum flash write IOPS for every second of every day for years. Applications also do many reads and have periods of high and low activity, such as day versus night, quarter close, end of a trading day, and so on. A very highly write intensive application might average 25 percent of the maximum flash write IOPS when measured over many months. Each Exadata X3 storage server has a total flash write endurance of over 50 PB for typical database data. In a full rack, if the application writes an average of 250,000 8K flash IOPS (25 percent of maximum writes) for 10 years, then it will write a total of 41 PB of data to each cell. This is less than the 50 PB per cell endurance.
If the ESM does not have sufficient charge, then the F20 PCIe card operates in fail-safe write-through mode, bypassing the DRAM memory and writing all data directly to flash. This results in reduced write performance, but there is no data loss. Exadata Storage Server generates an alert when the ESM capacity is insufficient, and the ESM should be replaced.
The charge capacity of the ESM degrades over time, and its life expectancy is inversely proportional to the operating temperature. The worst case life expectancy of the ESM in Oracle Exadata Rack is as follows:
Type of Exadata Storage Server | Lifetime |
---|---|
Exadata Storage Server with Sun Fire X4275 Servers |
3 years |
Exadata Storage Server with Sun Fire X4270 M2 Servers |
4 years |
Related Topics
Parent topic: Managing F20 PCIe Energy Storage Modules
3.11.2 Replacing Flash ESM
If the charge capacity in the F20 PCIe ESM falls below the minimum threshold, then Oracle will replace the failed ESM modules at no extra charge, if the system is covered either by the Oracle Premier Support for Systems or occurs during the warranty period.
For customers with Premier Support for Systems, Oracle attempts to proactively replace the F20 PCIe ESM in the Oracle Exadata Rack before the end of the estimated lifetime, on a best efforts basis.
Parent topic: Managing F20 PCIe Energy Storage Modules
3.12 Exadata Storage Server LED Indicator Descriptions
The indicator LEDs on Oracle Exadata storage servers help you to verify the system status and identify components that require servicing.
For information about the various indicator LEDs on Oracle Exadata storage servers, see the section on Troubleshooting Using the Server Front and Back Panel Status Indicators in the server service manual for your system.
See Related Documentation for a list of the server service manuals.
Additionally, the Do Not Service LED is included only on Oracle Exadata X7-2 and later storage servers.
On Oracle Exadata storage servers, the Do Not Service LED has the following states:
-
Do Not Service LED is white/on: Indicates that the storage server is required to remain online to preserve data availability. Do not restart or power off the storage server. Otherwise, data availability may be compromised.
Typically, the Do Not Service LED lights in response to an issue with a partner storage server. However, the Do Not Service LED also lights in the following situations:
-
During an Oracle ASM cluster rolling upgrade, the Do Not Service LED lights simultaneously on all participating storage servers. Furthermore, on affected grid disks, the
asmDeactivationOutcome
attribute contains the value:Cannot deactivate because ASM is in rolling upgrade mode
. -
When a database server containing a voting disk goes down, the Do Not Service LED lights simultaneously on all storage servers, which warns against shutting down any storage servers to preserve quorum in the cluster.
-
-
Do Not Service LED is off: The storage server can be safely powered off for servicing.
Parent topic: Maintaining Oracle Exadata Storage Servers
3.13 Exadata Storage Server Images
The Exadata Storage Server models have different external layouts and physical appearance.
- Oracle Exadata Storage Server X9M-2 Extreme Flash Server Images
- Oracle Exadata Storage Server X9M-2 High Capacity Server Images
- Oracle Exadata Storage Server X9M-2 Extended Server Images
- Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and Extended (XT) Server Images
- Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash Server Images
- Oracle Exadata Storage Server X7-2 High Capacity Server Images
- Oracle Exadata Storage Server X7-2 Extreme Flash Server Images
- High Capacity Exadata Storage Server X6-2 Images
- Extreme Flash Exadata Storage Server X6-2 Images
- High Capacity Exadata Storage Server X5-2 Images
- Extreme Flash Exadata Storage Server X5-2 Images
- Exadata Storage Server X4-2L Images
- Exadata Storage Server X3-2L Images
- Exadata Storage Server with Sun Fire X4270 M2 Images
- Exadata Storage Server with Sun Fire X4275 Images
Parent topic: Maintaining Oracle Exadata Storage Servers
3.13.1 Oracle Exadata Storage Server X9M-2 Extreme Flash Server Images
The following figure shows the front view of the Oracle Exadata Storage Server X9M-2 Extreme Flash (EF) servers.
Figure 3-1 Front View of Oracle Exadata Storage Server X9M-2 Extreme Flash Server

Description of "Figure 3-1 Front View of Oracle Exadata Storage Server X9M-2 Extreme Flash Server"
The following figure shows the rear view of the Oracle Exadata Storage Server X9M-2 Extreme Flash servers.
Figure 3-2 Rear View of Oracle Exadata Storage Server X9M-2 Extreme Flash Server

Description of "Figure 3-2 Rear View of Oracle Exadata Storage Server X9M-2 Extreme Flash Server"
Parent topic: Exadata Storage Server Images
3.13.2 Oracle Exadata Storage Server X9M-2 High Capacity Server Images
The following figure shows the front view of the Oracle Exadata Storage Server X9M-2 High Capacity (HC) servers.
Figure 3-3 Front View of Oracle Exadata Storage Server X9M-2 High Capacity servers

Description of "Figure 3-3 Front View of Oracle Exadata Storage Server X9M-2 High Capacity servers "
The following figure shows the rear view of the Oracle Exadata Storage Server X9M-2 High Capacity servers.
Figure 3-4 Rear View of Oracle Exadata Storage Server X9M-2 High Capacity servers

Description of "Figure 3-4 Rear View of Oracle Exadata Storage Server X9M-2 High Capacity servers "
Parent topic: Exadata Storage Server Images
3.13.3 Oracle Exadata Storage Server X9M-2 Extended Server Images
The following figure shows the front view of the Oracle Exadata Storage Server X9M-2 Extended (XT) servers.
Figure 3-5 Front View of Oracle Exadata Storage Server X9M-2 Extended servers

Description of "Figure 3-5 Front View of Oracle Exadata Storage Server X9M-2 Extended servers"
The following figure shows the rear view of the Oracle Exadata Storage Server X9M-2 Extended servers.
Figure 3-6 Rear View of Oracle Exadata Storage Server X9M-2 Extended servers

Description of "Figure 3-6 Rear View of Oracle Exadata Storage Server X9M-2 Extended servers"
Parent topic: Exadata Storage Server Images
3.13.4 Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and Extended (XT) Server Images
The following figure shows the front view of the Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and XT servers.
Figure 3-7 Front View of Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and XT servers

Description of "Figure 3-7 Front View of Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and XT servers "
The following figure shows the rear view of the Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and XT servers.
Figure 3-8 Rear View of Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and XT servers

Description of "Figure 3-8 Rear View of Oracle Exadata Storage Server X8M-2 and X8-2 High Capacity and XT servers "
Parent topic: Exadata Storage Server Images
3.13.5 Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash Server Images
The front view of the Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash server is almost identical to the X7-2 server. The main difference is the product logo.
Figure 3-9 Front View of Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash Server

Description of "Figure 3-9 Front View of Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash Server"
The rear view of the Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash server is almost identical to the X7-2 server. The main difference is the product logo.
Figure 3-10 Rear View of Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash Server

Description of "Figure 3-10 Rear View of Oracle Exadata Storage Server X8M-2 and X8-2 Extreme Flash Server"
Parent topic: Exadata Storage Server Images
3.13.6 Oracle Exadata Storage Server X7-2 High Capacity Server Images
The following figure shows the front view of the Oracle Exadata Storage Server X7-2 High Capacity Server.
Figure 3-11 Front View of Oracle Exadata Storage Server X7-2 High Capacity Server

Description of "Figure 3-11 Front View of Oracle Exadata Storage Server X7-2 High Capacity Server "
The following figure shows the rear view of the Oracle Exadata Storage Server X7-2 High Capacity Server.
Figure 3-12 Rear View of Oracle Exadata Storage Server X7-2 High Capacity Server

Description of "Figure 3-12 Rear View of Oracle Exadata Storage Server X7-2 High Capacity Server"
Parent topic: Exadata Storage Server Images
3.13.7 Oracle Exadata Storage Server X7-2 Extreme Flash Server Images
The following figure shows the front view of the Oracle Exadata Storage Server X7-2 Extreme Flash server.
Figure 3-13 Front View of Oracle Exadata Storage Server X7-2 Extreme Flash Server

Description of "Figure 3-13 Front View of Oracle Exadata Storage Server X7-2 Extreme Flash Server"
The following figure shows the rear view of the Oracle Exadata Storage Server X7-2 Extreme Flash server.
Figure 3-14 Rear View of Oracle Exadata Storage Server X7-2 Extreme Flash Server

Description of "Figure 3-14 Rear View of Oracle Exadata Storage Server X7-2 Extreme Flash Server"
Parent topic: Exadata Storage Server Images
3.13.8 High Capacity Exadata Storage Server X6-2 Images
The following figure shows the front view of the Oracle Exadata Storage Server X6-2 High Capacity server.
Figure 3-15 Front View of Oracle Exadata Storage Server X6-2 High Capacity Server

Description of "Figure 3-15 Front View of Oracle Exadata Storage Server X6-2 High Capacity Server "
The following figure shows the rear view of the Oracle Exadata Storage Server X6-2 High Capacity server.
Figure 3-16 Rear View of Oracle Exadata Storage Server X6-2 High Capacity Server

Description of "Figure 3-16 Rear View of Oracle Exadata Storage Server X6-2 High Capacity Server "
Parent topic: Exadata Storage Server Images
3.13.9 Extreme Flash Exadata Storage Server X6-2 Images
The following figure shows the front view of the Oracle Exadata Storage Server X6-2 Extreme Flash server.
Figure 3-17 Front View of Oracle Exadata Storage Server X6-2 Extreme Flash Server

Description of "Figure 3-17 Front View of Oracle Exadata Storage Server X6-2 Extreme Flash Server"
The following figure shows the rear view of the Oracle Exadata Storage Server X6-2 Extreme Flash server.
Figure 3-18 Rear View of Oracle Exadata Storage Server X6-2 Extreme Flash Server

Description of "Figure 3-18 Rear View of Oracle Exadata Storage Server X6-2 Extreme Flash Server "
Parent topic: Exadata Storage Server Images
3.13.10 High Capacity Exadata Storage Server X5-2 Images
The following image shows the front view of High Capacity Exadata Storage Server X5-2 Servers.
Figure 3-19 Front View of High Capacity Exadata Storage Server X5-2 Servers

Description of "Figure 3-19 Front View of High Capacity Exadata Storage Server X5-2 Servers"
The following image shows the rear view of High Capacity Exadata Storage Server X5-2 Servers.
Figure 3-20 Rear View of High Capacity Exadata Storage Server X5-2 Servers

Description of "Figure 3-20 Rear View of High Capacity Exadata Storage Server X5-2 Servers"
Parent topic: Exadata Storage Server Images
3.13.11 Extreme Flash Exadata Storage Server X5-2 Images
The following image shows the front view of Extreme Flash Exadata Storage Server X5-2 Servers.
Figure 3-21 Front View of Extreme Flash Exadata Storage Server X5-2 Servers

Description of "Figure 3-21 Front View of Extreme Flash Exadata Storage Server X5-2 Servers"
The following image shows the rear view of Extreme Flash Exadata Storage Server X5-2 Servers.
Figure 3-22 Rear View of Extreme Flash Exadata Storage Server X5-2 Servers

Description of "Figure 3-22 Rear View of Extreme Flash Exadata Storage Server X5-2 Servers"
Parent topic: Exadata Storage Server Images
3.13.12 Exadata Storage Server X4-2L Images
The following image shows the front view of Exadata Storage Server X4-2L Servers. The hard drives are numbered from left to right, starting in the lower left. The drives in the bottom row are numbers 0, 1, 2, and 3. The drives in the middle row are numbers 4, 5, 6, and 7. The drives in the top row are numbers 8, 9, 10, and 11.
Figure 3-23 Front View of Exadata Storage Server X4-2L Servers

Description of "Figure 3-23 Front View of Exadata Storage Server X4-2L Servers"
The following image shows the rear view of Exadata Storage Server X4-2L Servers.
Figure 3-24 Rear View of Exadata Storage Server X4-2L Servers

Description of "Figure 3-24 Rear View of Exadata Storage Server X4-2L Servers"
Parent topic: Exadata Storage Server Images
3.13.13 Exadata Storage Server X3-2L Images
The following image shows the front view of Exadata Storage Server X3-2L Servers. The hard drives are numbered from left to right, starting in the lower left. The drives in the bottom row are numbers 0, 1, 2, and 3. The drives in the middle row are numbers 4, 5, 6, and 7. The drives in the top row are numbers 8, 9, 10, and 11.
Figure 3-25 Front View of Exadata Storage Server X3-2L Servers

Description of "Figure 3-25 Front View of Exadata Storage Server X3-2L Servers"
The following image shows the rear view of Exadata Storage Server X3-2L Servers.
Figure 3-26 Rear View of Exadata Storage Server X3-2L Servers

Description of "Figure 3-26 Rear View of Exadata Storage Server X3-2L Servers"
Parent topic: Exadata Storage Server Images
3.13.14 Exadata Storage Server with Sun Fire X4270 M2 Images
The following image shows the front view of Exadata Storage Server with Sun Fire X4270 M2 Servers.
Figure 3-27 Front View of Exadata Storage Server with Sun Fire X4270 M2 Servers

Description of "Figure 3-27 Front View of Exadata Storage Server with Sun Fire X4270 M2 Servers"
-
Hard disk drives. The top drives are, from left to right, HDD2, HDD5, HDD8, and HDD11. The middle drives are, from left to right, HDD1, HDD4, HDD7, and HDD10. The bottom drives are, from left to right, HDD0, HDD3, HDD6, and HDD9.
The following image shows the rear view of Exadata Storage Server with Sun Fire X4270 M2 Servers.
Figure 3-28 Rear View of Exadata Storage Server with Sun Fire X4270 M2 Servers

Description of "Figure 3-28 Rear View of Exadata Storage Server with Sun Fire X4270 M2 Servers"
-
Power supplies.
-
InfiniBand host channel adapter PCI Express module.
-
Sun Flash Accelerator F20 PCIe Cards.
Parent topic: Exadata Storage Server Images
3.13.15 Exadata Storage Server with Sun Fire X4275 Images
The following figure shows the front view of Sun Fire X4275 Servers.
Figure 3-29 Front View of Sun Fire X4275 Server

Description of "Figure 3-29 Front View of Sun Fire X4275 Server"
-
Hard disk drives. The top drives are, from left to right, HDD2, HDD5, HDD8, and HDD11. The middle drives are, from left to right, HDD1, HDD4, HDD7, and HDD10. The bottom drives are, from left to right, HDD0, HDD3, HDD6, and HDD9.
The following figure shows the rear view of Sun Fire X4275 servers.
Figure 3-30 Rear View of Sun Fire X4275 Server

Description of "Figure 3-30 Rear View of Sun Fire X4275 Server"
Parent topic: Exadata Storage Server Images