This chapter describes the requirements and guidelines that are necessary to create RAID-1 volumes with the custom JumpStart or Solaris Live Upgrade installation methods.
This chapter describes the following topics.
For additional information about planning to create mirrored file systems with the Solaris Live Upgrade installation method, see General Guidelines When Creating RAID-1 Volumes (Mirrored) File Systems in Solaris 10 6/06 Installation Guide: Solaris Live Upgrade and Upgrade Planning.
For instructions about how to create mirrored file systems with the custom JumpStart installation method, see filesys Profile Keyword (Creating RAID-1 Volumes) and metadb Profile Keyword (Creating State Database Replicas).
To create RAID-1 volumes to duplicate data on specific slices, the disks that you plan to use must be directly attached and available to the system during the installation.
You should distribute state database replicas across slices, drives, and controllers, to avoid single points of failure. You want a majority of replicas to survive a single component failure. If you lose a replica, when a device fails, for example, the failure might cause problems with running Solaris Volume Manager software or when rebooting the system. Solaris Volume Manager software requires at least half of the replicas to be available to run, but a majority (half plus one) to reboot into multiuser mode.
For detailed instructions about creating and administering state database replicas, see Solaris Volume Manager Administration Guide.
Before selecting slices for state database replicas, consider the following guidelines and recommendations.
You should create state database replicas on a dedicated slice of at least 4 Mbytes per replica. If necessary, you could create state database replicas on a slice that is to be used as part of a RAID-0 or RAID-1 volume. You must create the replicas before you add the slice to the volume.
By default, the size of a state database replica is 4 Mbytes or 8192 disk blocks. Because your disk slices might not be that small, you can resize a slice to hold the state database replica. For information about resizing a slice, see Chapter 12, Administering Disks (Tasks), in System Administration Guide: Devices and File Systems.
You can create state database replicas on slices that are not in use. The part of a slice that is reserved for the state database replica should not be used for any other purpose.
You cannot create state database replicas on existing file systems, or the root (/), /usr, and swap file systems. If necessary, you can create a new slice (provided a slice name is available) by allocating space from swap and then put state database replicas on that new slice.
When a state database replica is placed on a slice that becomes part of a volume, the capacity of the volume is reduced by the space that is occupied by the replica or replicas. The space that is used by a replica is rounded up to the next cylinder boundary and this space is skipped by the volume.
Before choosing the number of state database replicas, consider the following guidelines.
A minimum of 3 state database replicas are recommended, up to a maximum of 50 replicas per Solaris Volume Manager disk set. The following guidelines are recommended:
For a system with only a single drive: put all three replicas in one slice.
For a system with two to four drives: put two replicas on each drive.
For a system with five or more drives: put one replica on each drive.
Additional state database replicas can improve the mirror's performance. Generally, you need to add two replicas for each mirror you add to the system.
If you have a RAID-1 volume that is to be used for small-sized random I/O (for example, for a database), consider your number of replicas. For best performance, ensure that you have at least two extra replicas per RAID-1 volume on slices (and preferably on disks and controllers) that are unconnected to the RAID-1 volume.
If multiple controllers exist, replicas should be distributed as evenly as possible across all controllers. This strategy provides redundancy if a controller fails and also helps balance the load. If multiple disks exist on a controller, at least two of the disks on each controller should store a replica.
When you are working with RAID-1 volumes (mirrors) and RAID-0 volumes (single-slice concatenations), consider the following guidelines.
The custom JumpStart installation method and Solaris Live Upgrade support a subset of the features that are available in the Solaris Volume Manager software. When you create mirrored file systems with these installation programs, consider the following guidelines.
Installation Program |
Supported Feature |
Unsupported Feature |
---|---|---|
Custom JumpStart and Solaris Live Upgrade |
|
In Solaris Volume manager a RAID-0 volume can refer to disk stripes or disk concatenations. You cannot create RAID-0 stripe volumes during the installation or upgrade. |
Custom JumpStart |
|
|
Solaris Live Upgrade |
For examples, see To Create a Boot Environment With RAID-1 Volumes (Mirrors) (Command-Line Interface) in Solaris 10 6/06 Installation Guide: Solaris Live Upgrade and Upgrade Planning. |
More than three RAID-0 volumes are not supported. |
Creating and Installing a Solaris Flash with RAID-1 volumes |
You can create a Solaris Flash archive created from a master system that has Solaris Volume Manager RAID-1 volumes configured. The Solaris Flash creation software removes all RAID-1 volume information from the archive to keep the integrity of the clone system. With custom JumpStart you can rebuild the RAID-1 volumes by using a JumpStart profile. With Solaris Live Upgrade, you create a boot environment with RAID-1 volumes configured and install the archive. The Solaris installation program cannot be used to install RAID-1 volumes with a Solaris Flash archive. For examples of RAID-1 volumes in JumpStart profiles, see Profile Examples. |
Veritas VxVM stores configuration information in areas not available to Solaris Flash. If Veritas VxVm file systems have been configured, you should not create a Solaris Flash archive. Also, Solaris install, including JumpStart and Solaris Live Upgrade do not support rebuilding VxVM volumes at installation time. Therefore, if you are planning to deploy Veritas VxVM software using a Solaris Flash archive, the archive must be created prior to configuring the VxVM file systems. The clone systems must be then configured individually after the archive has been applied and the system rebooted. |
Observe the following rules when assigning names for volumes.
Use a naming method that maps the slice number and disk number to volume numbers.
Volume names must begin with the letter d followed by a number, for example, d0.
Solaris Volume Manager has 128 default volume names from 0–127. The following list shows some example volume names.
Device /dev/md/dsk/d0 – block volume d0
Device /dev/md/dsk/d1 – block volume d1
Use ranges for each particular type of volume. For example, assign numbers 0–20 for RAID-1 volumes, and 21–40 for RAID-0 volumes.
Instead of specifying the full volume name, such as /dev/md/dsk/d1, you can often use an abbreviated volume name, such as d1.
You can abbreviate the names of physical disk slices and Solaris Volume Manager volumes. The abbreviation is the shortest name that uniquely identifies a device. Examples follow.
A Solaris Volume Manager volume can be identified by its dnum designation, so that, for example, /dev/md/dsk/d10 becomes simply d10.
If a system has a single controller and multiple disks, you might use t0d0s0, but with multiple controllers use c0t0d0s0.
When you use the Solaris Live Upgrade to create RAID-1 volumes (mirrors) and RAID-0 volumes (submirrors), you can let the software detect and assign volume names, or you can assign the names. If you let the software detect the names, the software assigns the first mirror or submirror name that is available. If you assign mirror names, assign names ending in zero so that the installation can use the names ending in 1 and 2 for submirrors. If you assign submirror names, assign names ending in 1 or 2. If you assign numbers incorrectly, the mirror might not be created. For example, if you specify a mirror name with a number that ends in 1 or 2 (d1 or d2), Solaris Live Upgrade fails to create the mirror if the mirror name is a duplicate of a submirror's name.
In this example, Solaris Live Upgrade assigns the volume names. The RAID-1 volumes d0 and d1 are the only volumes in use. For the mirror d10, Solaris Live Upgrade chooses d2 for the submirror for the device c0t0d0s0 and d3 for the submirror for the device c1t0d0s0.
lucreate -n newbe -m /:d10:mirror,ufs -m /:c0t0d0s0:attach -m /:c1t0d0s0:attach |
In this example, the volume names are assigned in the command. For the mirror d10, d11 is the name for the submirror for the device c0t0d0s0 and d12 is the name for the submirror for the device c1t0d0s0.
lucreate -n newbe -m /:d10:mirror,ufs -m /:c0t0d0s0,d11:attach -m /:c1t0d0s0,d12:attach |
For detailed information about Solaris Volume Manager naming requirements, see Solaris Volume Manager Administration Guide.
When you use the custom JumpStart installation method to create RAID-1 volumes (mirrors) and RAID-0 volumes (submirrors), you can let the software detect and assign volume names to mirrors, or you can assign the names in the profile. If you let the software detect the names, the software assigns the first volume number that is available. If you assign names in the profile, assign mirror names ending in zero so that the installation can use the names ending in 1 and 2 for submirrors. If you assign numbers incorrectly, the mirror might not be created. For example, if you specify a mirror name with a number that ends in 1 or 2 (d1 or d2), JumpStart fails to create the mirror if the mirror name is a duplicate of a submirror's name. In the following profile example, the mirror is assigned the first volume numbers that are available. If the next available mirror ending in zero is d10, then the names d11 and d12 are assigned to the submirrors.
filesys mirror c0t0d0s1 /
In the following profile example, the mirror number is assigned in the profile as d30. The submirror names are assigned by the software, based on the mirror number and the first available submirrors. In this example, the submirrors are named d31 and d32.
filesys mirror:d30 c0t1d0s0 c0t0d0s0 /
For detailed information about Solaris Volume Manager naming requirements, see Solaris Volume Manager Administration Guide.
When you choose the disks and controllers that you want to use to mirror a file system, consider the following guidelines.
Use components that are on different controllers to increase the number of simultaneous reads and writes that can be performed.
Keep the slices of different submirrors on different disks and controllers. Data protection is diminished considerably if slices of two or more submirrors of the same mirror are on the same disk.
Organize submirrors across separate controllers, because controllers and associated cables tend to fail more often than disks. This practice also improves mirror performance.
Use the same type of disks and controllers in a single mirror. Particularly in old SCSI storage devices, different models or brands of disk or controller can have widely varying performance. Mixing the different performance levels in a single mirror can cause performance to degrade significantly.
When you choose the slices that you want to use to mirror a file system, consider the following guidelines.
Any file system, including root (/), swap, and /usr, can use a mirror. Any application, such as a database, also can use a mirror.
Make sure that your submirror slices are of equal size. Submirrors of different sizes result in unused disk space.
If you have a mirrored file system in which the first submirror attached does not start on cylinder 0, all additional submirrors you attach must also not start on cylinder 0. If you attempt to attach a submirror starting on cylinder 0 to a mirror in which the original submirror does not start on cylinder 0, the following error message is displayed:
can't attach labeled submirror to an unlabeled mirror |
You must ensure that all submirrors you plan to attach to a mirror either all start on cylinder 0, or that none of them start on cylinder 0.
Starting cylinders do not have to be identical across all submirrors, but all submirrors must either include or not include cylinder 0.
If a system with mirrors for root (/), /usr, and swap is booted into single-user mode, the system indicates that these mirrors are in need of maintenance. When you view these mirrors with the metastat command, these mirrors, and possibly all mirrors on the system, appear in the “Needing Maintenance” state.
Though this situation appears to be potentially dangerous, do not be concerned. The metasync -r command, which normally occurs during boot to resynchronize mirrors, is interrupted when the system is booted into single-user mode. After the system is rebooted, the metasync -r command runs and resynchronizes all mirrors.
If this interruption is a concern, run the metasync -r command manually.
For more information about the metasync, see the metasync(1M) man page, and Solaris Volume Manager Administration Guide.