Skip Navigation Links | |
Exit Print View | |
Solaris Volume Manager Administration Guide Oracle Solaris 10 1/13 Information Library |
1. Getting Started With Solaris Volume Manager
2. Storage Management Concepts
3. Solaris Volume Manager Overview
4. Solaris Volume Manager for Sun Cluster (Overview)
5. Configuring and Using Solaris Volume Manager (Scenario)
8. RAID-0 (Stripe and Concatenation) Volumes (Overview)
9. RAID-0 (Stripe and Concatenation) Volumes (Tasks)
10. RAID-1 (Mirror) Volumes (Overview)
11. RAID-1 (Mirror) Volumes (Tasks)
12. Soft Partitions (Overview)
16. Hot Spare Pools (Overview)
20. Maintaining Solaris Volume Manager (Tasks)
21. Best Practices for Solaris Volume Manager
22. Top-Down Volume Creation (Overview)
23. Top-Down Volume Creation (Tasks)
24. Monitoring and Error Reporting (Tasks)
25. Troubleshooting Solaris Volume Manager (Tasks)
Troubleshooting Solaris Volume Manager (Task Map)
Overview of Troubleshooting the System
Prerequisites for Troubleshooting the System
General Guidelines for Troubleshooting Solaris Volume Manager
General Troubleshooting Approach
Device ID Discrepancies After Upgrading to the Solaris 10 Release
Background Information for Boot Problems
How to Recover From Improper /etc/vfstab Entries
Recovering the root (/) RAID-1 (Mirror) Volume
How to Recover From a Boot Device Failure
Recovering From State Database Replica Failures
How to Recover From Insufficient State Database Replicas
Recovering From Soft Partition Problems
How to Recover Configuration Data for a Soft Partition
Recovering Storage From a Different System
How to Recover Storage From a Local Disk Set
Recovering Storage From a Known Disk Set
How to Print a Report on Disk Sets Available for Import
How to Import a Disk Set From One System to Another System
Recovering From Disk Set Problems
What to Do When You Cannot Take Ownership of A Disk Set
Performing Mounted Filesystem Backups Using the ufsdump Command
How to Perform a Backup of a Mounted Filesystem Located on a RAID-1 Volume
How to Recover a System Using a Solaris Volume Manager Configuration
A. Important Solaris Volume Manager Files
B. Solaris Volume Manager Quick Reference
This section describes how to recover from unexpected problems after moving disks in the Solaris Volume Manager environment.
Solaris Volume Manager uses device IDs, which are associated with a specific disk, to track all disks that are used in a Solaris Volume Manager configuration. When disks are moved to a different controller or when the SCSI target numbers change, Solaris Volume Manager usually correctly identifies the movement and updates all related Solaris Volume Manager records accordingly. No system administrator intervention is required. In isolated cases, Solaris Volume Manager cannot completely update the records and reports an error on boot.
If you add new hardware or move hardware (for example, you move a string of disks from one controller to another controller), Solaris Volume Manager checks the device IDs that are associated with the disks that moved, and updates the cntndn names in internal Solaris Volume Manager records accordingly. If the records cannot be updated, the boot processes that are spawned by the svc:/system/mdmonitor service report an error to the console at boot time:
Unable to resolve unnamed devices for volume management. Please refer to the Solaris Volume Manager documentation, Troubleshooting section, at http://docs.sun.com or from your local copy.
No data loss has occurred, and none will occur as a direct result of this problem. This error message indicates that the Solaris Volume Manager name records have been only partially updated. Output from the metastat command shows some of the cntndn names that were previously used. The output also shows some of the cntndn names that reflect the state after the move.
If you need to update your Solaris Volume Manager configuration while this condition exists, you must use the cntndn names that are reported by the metastat command when you issue any meta* commands.
If this error condition occurs, you can do one of the following to resolve the condition:
Restore all disks to their original locations. Next, do a reconfiguration reboot, or run (as a single command):
/usr/sbin/devfsadm && /usr/sbin/metadevadm -r
After these commands complete, the error condition is resolved.
Contact your support representative for guidance.
Note - This error condition is quite unlikely to occur. If it does occur, it is most likely to affect Fibre Channel-attached storage.