JavaScript is required to for searching.
Skip Navigation Links
Exit Print View
Oracle VM Server for SPARC 2.0 Release Notes
search filter icon
search icon

Document Information

Preface

1.  Oracle VM Server for SPARC 2.0 Release Notes

What's New in This Release

System Requirements

Supported Platforms

Required Software and Patches

Required Logical Domains Manager

Required and Recommended Oracle Solaris OS

Required Software to Enable Oracle VM Server for SPARC 2.0 Features

Required and Recommended System Firmware Patches

Minimum Version of Software Required

Direct I/O Hardware and Software Requirements

Direct I/O Hardware Requirements

Direct I/O Software Requirements

Location of Oracle VM Server for SPARC 2.0 Software

Location of Patches

Location of Documentation

Related Software

Optional Software

Software That Can Be Used With the Logical Domains Manager

System Controller Software That Interacts With Logical Domains Software

Upgrading to Oracle VM Server for SPARC 2.0 Software

Upgrade From Logical Domains 1.0 Software Only

Known Issues

General Issues

I/O MMU Bypass Mode Is No Longer Needed

Service Processor and System Controller Are Interchangeable Terms

Cards Not Supported

In Certain Conditions, a Guest Domain's Solaris Volume Manager Configuration or Metadevices Can Be Lost

Logical Domain Channels and Logical Domains

Memory Size Requirements

Booting a Large Number of Domains

Cleanly Shutting Down and Power Cycling a Logical Domains System

Memory Size Requested Might Be Different From Memory Allocated

Logical Domains Variable Persistence

Oracle's Sun SNMP Management Agent Does Not Support Multiple Domains

Containers, Processor Sets, and Pools Are Not Compatible With CPU Power Management

Fault Management

Delayed Reconfiguration

Cryptographic Units

Bugs Affecting the Oracle VM Server for SPARC 2.0 Software

Logical Domains Manager Does Not Validate Disk Paths and Network Devices

Disk Paths

Network Devices

Hang Can Occur With Guest OS in Simultaneous Operations

Behavior of the ldm stop-domain Command Can Be Confusing

Cannot Set Security Keys With Logical Domains Running

OpenBoot PROM Variables Cannot be Modified by the eeprom(1M) Command When the Logical Domains Manager is Running

Logical Domains Manager Does Not Retire Resources On Guest Domain After a Panic and Reboot

Guest Domain With Too Many Virtual Networks on the Same Network Using DHCP Can Become Unresponsive

The scadm Command Can Hang Following an SC or SP Reset

Simultaneous Net-Installation of Multiple Domains Fails When in a Common Console Group

Solaris Volume Manager Volumes Built on Slice 2 Fail JumpStart When Used as the Boot Device in a Guest Domain

If the Oracle Solaris 10 5/08 OS Is Installed on a Service Domain, Attempting a Net Boot of the Oracle Solaris 10 8/07 OS on Any Guest Domain Serviced by It Can Hang the Installation

Logical Domains Manager Can Take Over 15 Minutes to Shut Down a Domain

Sometimes, Executing the uadmin 1 0 Command From an Logical Domains System Does Not Return the System to the OK Prompt

Logical Domains Manager Displays Migrated Domains in Transition States When They Are Already Booted

Logical Domains Manager Does Not Start If the Machine Is Not Networked and an NIS Client Is Running

Migration Can Fail to Bind Memory Even If the Target Has Enough Available

Migration Does Not Fail If a vdsdev on the Target Has a Different Back End

Constraint Database Is Not Synchronized to Saved Configuration

Explicit Console Group and Port Bindings Are Not Migrated

VIO DR Operations Ignore the Force (-f) Option

ldmconfig Might Cause the Root File System of the Control Domain to Become Full and Halt the System

Guest Domain Sometimes Fails to Make Proper Domain Services Connection to the Control Domain

ldmd Might Dump Core If Multiple set-vcpu Operations Are Performed on the Control Domain While It Is in Delayed Reconfiguration Mode

Virtual Network Devices Are Not Created Properly on the Control Domain

Cannot Connect to Migrated Domain's Console Unless vntsd Is Restarted

I/O Domain or Guest Domain Panics When Booting From e1000g

Guest Domain Might Fail to Successfully Reboot When a System Is in Power Management Elastic Mode

ldm Commands Are Slow to Respond When Several Domains Are Booting

Spurious ds_ldc_cb: LDC READ event Message Seen When Rebooting the Control Domain or a Guest Domain

Migrated Domain With MAUs Contains Only One CPU When Target OS Does Not Support DR of Cryptographic Units

Confusing Migration Failure Message for Real Address Memory Bind Failures

Newly Added NIU/XAUI Adapters Are Not Visible to Host OS If Logical Domains Is Configured

Dynamically Removing All the Cryptographic Units From a Domain Causes SSH to Terminate

Atlas PCI Express Dual 10-Gigabit Ethernet Fiber Card Shows Four Subdevices in ldm list-io -l Output

ldm set-domain -i Command Fails to Completely Configure the Domain as Specified by the XML File

ldmp2v Command: ufsdump Archiving Method Is No Longer Used

ldmp2v convert Command: VxVM Warning Messages During Boot

Dynamic Reconfiguration of MTU Values of Virtual Network Devices Sometimes Fails

Active Guest Domain Erroneously Shown to Be in Transition Mode

ldmconfig -c Command Fails Silently

Memory DR Is Disabled Following a Canceled Migration

Migration of a Guest Domain That Has Hybrid I/O-Enabled Virtual Network Devices Panics the Service Domain

Migrated Guest Domain Hangs If It Has CPUs That Are Offline, Faulted, or Spare

ldmconfig Command Does Not Save Virtual Disk Service Devices for Bound Domains

Migration of a Domain That Has an Enabled Default DRM Policy Results in a Target Domain Being Assigned All Available CPUs

Uncooperative Oracle Solaris Domain Migration Can Be Blocked If cpu0 Is Offline

Sometimes DR Requests Fail to Remove All Requested CPUs

When Re-creating a Guest Domain From an XML File, the Domain Should be Created by Using the CPU Allocation Unit and Not the Whole Core Number

Inactive or Bound Domains Allow the Coexistence of the Whole-Core Constraint With an Enabled DRM Policy

ldm set-vcc port-range Command While in Delayed Reconfiguration Mode Fails to Hold the New port-range Value After Reboot

Using ldm stop -a Command on Domains in a Master-Slave Relationship Leaves the Slave With the stopping Flag Set

Guest Domain Might Panic After the Domain Is Started With the ldm start Command

ldmp2v convert Command: The drd Daemon Is Not Always Enabled

ldc_close: (0xb) unregister failed, 11 Warning Messages

Memory DR Removal Operations With Unplumbed NIU nxge Instances Can Cause a Kernel Panic, Panic - Generated at user request

Memory DR Removal Operations With Multiple Plumbed NIU nxge Instances Can Hang Indefinitely and Never Complete

Direct I/O Configuration Is Not Properly Rebuilt by the ldm init-system Command

Guest Domains Cannot Boot From an Exported DVD Device

Emulex HBA Permanently Destroyed If the Firmware Upgrade Is Interrupted

Block of Dynamically Added Memory Can Only Be Dynamically Removed as a Whole

Cryptographic Units Cannot Be Removed From the primary Domain

SPARC T3-1: Detect And Handle Disks That Are Accessible Through Multiple Direct I/O Paths

System Powers Off When a Memory Request Results in a Hypervisor Abort

nxge Driver Supports Only Eight Hybrid I/O Shares in a Guest Domain

Documentation Errata

Incorrect Parameter Names in the Input/Output Bus Table

virtinfo(1M) and Related API Man Pages Missing From the Oracle Solaris 10 9/10 Release

Cannot Upgrade From Logical Domains 1.0 to Oracle VM Server for SPARC 2.0

Resolved Issues

Oracle VM Server for SPARC 2.0 RFEs and Bugs Fixed in Oracle Solaris 10 9/10

RFEs and Bugs Fixed for Oracle VM Server for SPARC 2.0 Software

Known Issues

This section contains general issues and specific bugs concerning the Oracle VM Server for SPARC 2.0 software.

General Issues

This section describes general known issues about this release of the Oracle VM Server for SPARC software that are broader than a specific bug number. Workarounds are provided where available.

I/O MMU Bypass Mode Is No Longer Needed

Starting with the Oracle VM Server for SPARC 2.0 release, I/O memory management unit (MMU) bypass mode is no longer needed. As a result, the bypass=on property is no longer available for use by the ldm add-io command.

Service Processor and System Controller Are Interchangeable Terms

For discussions in Oracle VM Server for SPARC documentation, the terms service processor (SP) and system controller (SC) are interchangeable.

Cards Not Supported

The following cards are not supported for this Oracle VM Server for SPARC 2.0 software release:


Caution

Caution - If these unsupported configurations are used with Oracle VM Server for SPARC 2.0, stop and unbind all logical domains before the control domain is rebooted. Failure to do so can result in a system crash causing the loss of all the logical domains that are active in the system.


In Certain Conditions, a Guest Domain's Solaris Volume Manager Configuration or Metadevices Can Be Lost

If a service domain is running a version of Oracle Solaris 10 OS prior to Oracle Solaris 10 9/10 and is exporting a physical disk slice as a virtual disk to a guest domain, then this virtual disk will appear in the guest domain with an inappropriate device ID. If that service domain is then upgraded to Oracle Solaris 10 9/10, the physical disk slice exported as a virtual disk will appear in the guest domain with no device ID.

This removal of the device ID of the virtual disk can cause problems to applications attempting to reference the device ID of virtual disks. In particular, this can cause the Solaris Volume Manager to be unable to find its configuration or to access its metadevices.

Workaround: After upgrading a service domain to Oracle Solaris 10 9/10, if a guest domain is unable to find its Solaris Volume Manager configuration or its metadevices, execute the following procedure.

Find a Guest Domain's Solaris Volume Manager Configuration or Metadevices

  1. Boot the guest domain.
  2. Disable the devid feature of Solaris Volume Manager by adding the following lines to the /kernel/dr/md.conf file:
    md_devid_destroy=1;
    md_keep_repl_state=1;
  3. Reboot the guest domain.

    After the domain has booted, the Solaris Volume Manager configuration and metadevices should be available.

  4. Check the Solaris Volume Manager configuration and ensure that it is correct.
  5. Re-enable the Solaris Volume Manager devid feature by removing from the /kernel/drv/md.conf file the two lines that you added in Step 2.
  6. Reboot the guest domain.

    During the reboot, you will see messages similar to this:

    NOTICE: mddb: unable to get devid for 'vdc', 0x10

    These messages are normal and do not report any problems.

Logical Domain Channels and Logical Domains

There is a limit to the number of logical domain channels (LDCs) that are available in any logical domain. For UltraSPARC T2 servers, SPARC T3-1 servers, and SPARC T3-1B servers, the limit is 512. For UltraSPARC T2 Plus servers and the other SPARC T3 servers, the limit is 768. This only becomes an issue on the control domain because the control domain has at least part, if not all, of the I/O subsystem allocated to it. This might also be an issue because of the potentially large number of LDCs that are created for both virtual I/O data communications and the Logical Domains Manager control of the other logical domains.

If you try to add a service, or bind a domain, so that the number of LDC channels exceeds the limit on the control domain, the operation fails with an error message similar to the following:

13 additional LDCs are required on guest primary to meet this request,
but only 9 LDCs are available

The following guidelines can help prevent creating a configuration that could overflow the LDC capabilities of the control domain:

  1. The control domain allocates 12 LDCs for various communication purposes with the hypervisor, Fault Management Architecture (FMA), and the system controller (SC), independent of the number of other logical domains configured.

  2. The control domain allocates 1 LDC to every logical domain, including itself, for control traffic.

  3. Each virtual I/O service on the control domain consumes 1 LDC for every connected client of that service.

For example, consider a control domain and 8 additional logical domains. Each logical domain needs the following at a minimum:

Applying the above guidelines yields the following results (numbers in parentheses correspond to the preceding guideline number from which the value was derived):

12(1) + 9(2) + 8 x 3(3)=45 LDCs in total.

Now consider the case where there are 45 domains instead of 8, and each domain includes 5 virtual disks, 5 virtual networks, and a virtual console. Now the equation becomes:

12 + 46 + 45 x 11=553 LDCs in total.

Depending upon the number of supported LDCs of your platform, the Logical Domains Manager will either accept or reject the configurations.

Memory Size Requirements

The Oracle VM Server for SPARC software does not impose a memory size limitation when you create a domain. The memory size requirement is a characteristic of the guest operating system. Some Oracle VM Server for SPARC functionality might not work if the amount of memory present is less than the recommended size. For recommended and minimum size memory requirements for the Oracle Solaris 10 OS, see System Requirements and Recommendations in Oracle Solaris 10 9/10 Installation Guide: Planning for Installation and Upgrade.

The OpenBoot PROM has a minimum size restriction for a domain. Currently, that restriction is 12 Mbytes. If you have a domain less than that size, the Logical Domains Manager will automatically boost the size of the domain to 12 Mbytes. Refer to the release notes for your system firmware for information about memory size requirements.

The memory dynamic reconfiguration (DR) feature enforces 256-Mbyte alignment on the address and size of the memory involved in a given operation. See Memory Alignment in Oracle VM Server for SPARC 2.0 Administration Guide.

Booting a Large Number of Domains

You can boot the following number of domains depending on your platform:

If unallocated virtual CPUs are available, assign them to the service domain to help process the virtual I/O requests. Allocate 4 to 8 virtual CPUs to the service domain when creating more than 32 domains. In cases where maximum domain configurations have only a single CPU in the service domain, do not put unnecessary stress on the single CPU when configuring and using the domain. The virtual switch (vsw) services should be spread over all the network adapters available in the machine. For example, if booting 128 domains on a Sun SPARC Enterprise T5240 server, create 4 vsw services, each serving 32 virtual net (vnet) instances. Do not have more than 32 vnet instances per vsw service because having more than that tied to a single vsw could cause hard hangs in the service domain.

To run the maximum configurations, a machine needs the following amount of memory, depending on your platform, so that the guest domains contain an adequate amount of memory:

Memory and swap space usage increases in a guest domain when the vsw services used by the domain provides services to many virtual networks (in multiple domains). This is due to the peer-to-peer links between all the vnet connected to the vsw. The service domain benefits from having extra memory. Four Gbytes is the recommended minimum when running more than 64 domains. Start domains in groups of 10 or less and wait for them to boot before starting the next batch. The same advice applies to installing operating systems on domains.

Cleanly Shutting Down and Power Cycling a Logical Domains System

If you have made any configuration changes since last saving a configuration to the SC, before you attempt to power off or power cycle a Logical Domains system, make sure that you save the latest configuration that you want to keep.

Power Off a System With Multiple Active Domains

  1. Shut down and unbind all the non-I/O domains.
  2. Shut down and unbind any active I/O domains.
  3. Halt the primary domain.

    Because no other domains are bound, the firmware automatically powers off the system.

Power Cycle the System

  1. Shut down and unbind all the non-I/O domains.
  2. Shut down and unbind any active I/O domains.
  3. Reboot the primary domain.

    Because no other domains are bound, the firmware automatically power cycles the system before rebooting it. When the system restarts, it boots into the Logical Domains configuration last saved or explicitly set.

Memory Size Requested Might Be Different From Memory Allocated

Under certain circumstances, the Logical Domains Manager rounds up the requested memory allocation to either the next largest 8-Kbyte or 4-Mbyte multiple. This can be seen in the following example output of the ldm list-domain -l command, where the constraint value is smaller than the actual allocated size:

Memory:
          Constraints: 1965 M
          raddr          paddr5          size
          0x1000000      0x291000000     1968M
Logical Domains Variable Persistence

Variable updates persist across a reboot, but not across a powercycle, unless the variable updates are either initiated from OpenBoot firmware on the control domain or followed by saving the configuration to the SC.

In this context, it is important to note that a reboot of the control domain could initiate a powercycle of the system:

Logical Domains variables for a domain can be specified using any of the following methods:

The goal is that, variable updates that are made by using any of these methods always persist across reboots of the domain. The variable updates also always reflect in any subsequent logical domain configurations that were saved to the SC.

In Oracle VM Server for SPARC 2.0 software, there are a few cases where variable updates do not persist as expected:

If you are concerned about Logical Domains variable changes, do one of the following:

If you modify the time or date on a logical domain, for example using the ntpdate command, the change persists across reboots of the domain but not across a power cycle of the host. To ensure that time changes persist, save the configuration with the time change to the SP and boot from that configuration.

The following Bug IDs have been filed to resolve these issues: 6520041, 6540368, 6540937, and 6590259.

Oracle's Sun SNMP Management Agent Does Not Support Multiple Domains

Sun Simple Management Network Protocol (SNMP) Management Agent does not support multiple domains. Only a single global domain is supported.

Containers, Processor Sets, and Pools Are Not Compatible With CPU Power Management

Using CPU dynamic reconfiguration (DR) to power down virtual CPUs does not work with processor sets, resource pools, or the zone's dedicated CPU feature.

When using CPU power management in elastic mode, the Oracle Solaris OS guest sees only the CPUs that are allocated to the domains that are powered on. That means that output from the psrinfo(1M) command dynamically changes depending on the number of CPUs currently power-managed. This causes an issue with processor sets and pools, which require actual CPU IDs to be static to allow allocation to their sets. This can also impact the zone's dedicated CPU feature.

Workaround: Set the performance mode for the power management policy.

Fault Management

There are several issues associated with FMA and power-managing CPUs. If a CPU faults when running in elastic mode, switch to performance mode until the faulted CPU recovers. If all faulted CPUs recover, then elastic mode can be used again.

Delayed Reconfiguration

When a primary domain is in a delayed reconfiguration state, CPUs are power managed only after the primary domain reboots. This means that CPU power management will not bring additional CPUs online while the domain is experiencing high-load usage until the primary domain reboots, clearing the delayed reconfiguration state.

Cryptographic Units

The Oracle Solaris 10 10/09 OS introduces the capability to dynamically add and remove cryptographic units from a domain, which is called cryptographic unit dynamic reconfiguration (DR). The Logical Domains Manager automatically detects whether a domain allows cryptographic unit DR, and enables the functionality only for those domains. In addition, CPU DR is no longer disabled in domains that have cryptographic units bound and are running an appropriate version of the Oracle Solaris OS.

No core disable operations are performed on domains that have cryptographic units bound when the SP is set to elastic mode. To enable core disable operations to be performed when the system is in elastic mode, remove the cryptographic units that are bound to the domain.

Bugs Affecting the Oracle VM Server for SPARC 2.0 Software

This section summarizes the bugs that you might encounter when using this version of the software. The bug descriptions are in numerical order by bug ID. If a workaround and a recovery procedure are available, they are specified.

Logical Domains Manager Does Not Validate Disk Paths and Network Devices

Bug ID 6447740: The Logical Domains Manager does not validate disk paths and network devices.

Disk Paths

If a disk device listed in a guest domain's configuration is either non-existent or otherwise unusable, the disk cannot be used by the virtual disk server (vds). However, the Logical Domains Manager does not emit any warning or error when the domain is bound or started.

When the guest tries to boot, messages similar to the following are printed on the guest's console:

WARNING: /virtual-devices@100/channel-devices@200/disk@0: Timeout
connecting to virtual disk server... retrying

In addition, if a network interface specified using the net-dev= property does not exist or is otherwise unusable, the virtual switch is unable to communicate outside the physical machine, but the Logical Domains Manager does not emit any warning or error when the domain is bound or started.

Recover From an Errant net-dev Property Specified for a Virtual Switch

  1. Issue the ldm set-vsw command with the corrected net-dev property value.
  2. Reboot the domain hosting the virtual switch in question.

Recover From an Errant Virtual Disk Service Device or Volume

  1. Stop the domain owning the virtual disk bound to the errant device or volume.
  2. Issue the ldm rm-vdsdev command to remove the errant virtual disk service device.
  3. Issue the ldm add-vdsdev command to correct the physical path to the volume.
  4. Restart the domain owning the virtual disk.
Network Devices

If a disk device listed in a guest domain's configuration is being used by software other than the Logical Domains Manager (for example, if it is mounted in the service domain), the disk cannot be used by the virtual disk server (vds), but the Logical Domains Manager does not emit a warning that it is in use when the domain is bound or started.

When the guest domain tries to boot, a message similar to the following is printed on the guest's console:

WARNING: /virtual-devices@100/channel-devices@200/disk@0: Timeout
connecting to virtual disk server... retrying

Recover From a Disk Device Being Used by Other Software

  1. Unbind the guest domain.
  2. Unmount the disk device to make it available.
  3. Bind the guest domain.
  4. Boot the domain.
Hang Can Occur With Guest OS in Simultaneous Operations

Bug ID 6497796: Under rare circumstances, when a Logical Domains variable, such as boot-device, is being updated from within a guest domain by using the eeprom(1M) command at the same time that the Logical Domains Manager is being used to add or remove virtual CPUs from the same domain, the guest OS can hang.

Workaround: Ensure that these two operations are not performed simultaneously.

Recovery: Use the ldm stop-domain and ldm start-domain commands to stop and start the guest OS.

Behavior of the ldm stop-domain Command Can Be Confusing

Bug ID 6506494: There are some cases where the behavior of the ldm stop-domain command is confusing.

# ldm stop-domain -f ldom

If the domain is at the kernel module debugger, kmdb(1), prompt, then the ldm stop-domain command fails with the following error message:

LDom <domain name> stop notification failed
Cannot Set Security Keys With Logical Domains Running

Bug ID 6510214: In a Logical Domains environment, there is no support for setting or deleting wide-area network (WAN) boot keys from within the Oracle Solaris OS by using the ickey(1M) command. All ickey operations fail with the following error:

ickey: setkey: ioctl: I/O error

In addition, WAN boot keys that are set using OpenBoot firmware in logical domains other than the control domain are not remembered across reboots of the domain. In these domains, the keys set from the OpenBoot firmware are only valid for a single use.

OpenBoot PROM Variables Cannot be Modified by the eeprom(1M) Command When the Logical Domains Manager is Running

Bug ID 6540368: This issue is summarized in Logical Domains Variable Persistence and affects only the control domain.

Logical Domains Manager Does Not Retire Resources On Guest Domain After a Panic and Reboot

Bug ID 6591844: If a CPU or memory fault occurs, the affected domain might panic and reboot. If the Fault Management Architecture (FMA) attempts to retire the faulted component while the domain is rebooting, the Logical Domains Manager is not able to communicate with the domain, and the retire fails. In this case, the fmadm faulty command lists the resource as degraded.

Recovery: Wait for the domain to complete rebooting, and then force FMA to replay the fault event by restarting the fault manager daemon (fmd) on the control domain by using this command:

primary# svcadm restart fmd
Guest Domain With Too Many Virtual Networks on the Same Network Using DHCP Can Become Unresponsive

Bug ID 6603974: If you configure more than four virtual networks (vnets) in a guest domain on the same network using the Dynamic Host Configuration Protocol (DHCP), the guest domain can eventually become unresponsive while running network traffic.

Workaround: Set ip_ire_min_bucket_cnt and ip_ire_max_bucket_cnt to larger values, such as 32, if you have 8 interfaces.

Recovery: Issue an ldm stop-domain ldom command followed by an ldm start-domain ldom command on the guest domain (ldom) in question.

The scadm Command Can Hang Following an SC or SP Reset

Bug ID 6629230: The scadm command on a control domain running at least the Solaris 10 11/06 OS can hang following an SC reset. The system is unable to properly reestablish a connection following an SC reset.

Workaround: Reboot the host to reestablish connection with the SC.

Recovery: Reboot the host to reestablish connection with the SC.

Simultaneous Net-Installation of Multiple Domains Fails When in a Common Console Group

Bug ID 6656033: Simultaneous net installation of multiple guest domains fails on Sun SPARC Enterprise T5140 and Sun SPARC Enterprise T5240 systems that have a common console group.

Workaround: Only net-install on guest domains that each have their own console group. This failure is seen only on domains with a common console group shared among multiple net-installing domains.

Solaris Volume Manager Volumes Built on Slice 2 Fail JumpStart When Used as the Boot Device in a Guest Domain

Bug ID 6687634: If the Solaris Volume Manager volume is built on top of a disk slice that contains block 0 of the disk, then Solaris Volume Manager prevents writing to block 0 of the volume to avoid overwriting the label of the disk.

If an Solaris Volume Manager volume built on top of a disk slice that contains block 0 of the disk is exported as a full virtual disk, then a guest domain is unable to write a disk label for that virtual disk, and this prevents the Oracle Solaris OS from being installed on such a disk.

Workaround: Solaris Volume Manager volumes exported as a virtual disk should not be built on top of a disk slice that contains block 0 of the disk.

A more generic guideline is that slices that start on the first block (block 0) of a physical disk should not be exported (either directly or indirectly) as a virtual disk. Refer to Directly or Indirectly Exporting a Disk Slice in Oracle VM Server for SPARC 2.0 Administration Guide.

If the Oracle Solaris 10 5/08 OS Is Installed on a Service Domain, Attempting a Net Boot of the Oracle Solaris 10 8/07 OS on Any Guest Domain Serviced by It Can Hang the Installation

Bug ID 6705823: Attempting a net boot of the Oracle Solaris 10 8/07 OS on any guest domain serviced by a service domain running the Oracle Solaris 10 5/08 OS can result in a hang on the guest domain during the installation.

Workaround: Patch the miniroot of the Oracle Solaris 10 8/07 OS net install image with Patch ID 127111-05.

Logical Domains Manager Can Take Over 15 Minutes to Shut Down a Domain

Bug ID 6742805: A domain shutdown or memory scrub can take over 15 minutes with a single CPU and a very large memory configuration. During a shutdown, the CPUs in a domain are used to scrub all the memory owned by the domain. The time taken to complete the scrub can be quite long if a configuration is imbalanced, for example, a single CPU domain with 512 Gbytes of memory. This prolonged scrub time extends the amount of time it takes to shut down a domain.

Workaround: Ensure that large memory configurations (>100 Gbytes) have at least one core. This results in a much faster shutdown time.

Sometimes, Executing the uadmin 1 0 Command From an Logical Domains System Does Not Return the System to the OK Prompt

Bug ID 6753683: Sometimes, executing the uadmin 1 0 command from the command line of an Logical Domains system does not leave the system at the ok prompt after the subsequent reset. This incorrect behavior is seen only when the Logical Domains variable auto-reboot? is set to true. If auto-reboot? is set to false, the expected behavior occurs.

Workaround: Use this command instead:

uadmin 2 0

Or, always run with auto-reboot? set to false.

Logical Domains Manager Displays Migrated Domains in Transition States When They Are Already Booted

Bug ID 6760933: On occasion, an active logical domain appears to be in the transition state instead of the normal state long after it is booted or following the completion of a domain migration. This glitch is harmless, and the domain is fully operational. To see what flag is set, check the flags field in the ldm list -l -p command output, or check the FLAGS field in the ldm list command, which shows -n---- for normal or -t---- for transition.

Recovery: The logical domain should display the correct state upon the next reboot.

Logical Domains Manager Does Not Start If the Machine Is Not Networked and an NIS Client Is Running

Bug ID 6764613: If you do not have a network configured on your machine and have a Network Information Services (NIS) client running, the Logical Domains Manager will not start on your system.

Workaround: Disable the NIS client on your non-networked machine:

# svcadm disable nis/client
Migration Can Fail to Bind Memory Even If the Target Has Enough Available

Bug ID 6772089: In certain situations, a migration fails and ldmd reports that it was not possible to bind the memory needed for the source domain. This can occur even if the total amount of available memory on the target machine is greater than the amount of memory being used by the source domain.

This failure occurs because migrating the specific memory ranges in use by the source domain requires that compatible memory ranges are available on the target, as well. When no such compatible memory range is found for any memory range in the source, the migration cannot proceed.

Recovery: If this condition is encountered, you might be able to migrate the domain if you modify the memory usage on the target machine. To do this, unbind any bound or active logical domain on the target.

Use the ldm list-devices -a mem command to see what memory is available and how it is used. You might also need to reduce the amount of memory that is assigned to another domain.

Migration Does Not Fail If a vdsdev on the Target Has a Different Back End

Bug ID 6772120: If the virtual disk on the target machine does not point to the same disk back end that is used on the source machine, the migrated domain cannot access the virtual disk using that disk back end. A hang can result when accessing the virtual disk on the domain.

Currently, the Logical Domains Manager checks only that the virtual disk volume names match on the source and target machines. In this scenario, no error message is displayed if the disk back ends do not match.

Workaround: Ensure that when you are configuring the target domain to receive a migrated domain that the disk volume (vdsdev) matches the disk back end used on the source domain.

Recovery: Do one of the following if you discover that the virtual disk device on the target machine points to the incorrect disk back end:

Constraint Database Is Not Synchronized to Saved Configuration

Bug ID 6773569: After switching from one configuration to another (using the ldm set-config command followed by a powercycle), domains defined in the previous configuration might still be present in the current configuration, in the inactive state.

This is a result of the Logical Domains Manager's constraint database not being kept in sync with the change in configuration. These inactive domains do not affect the running configuration and can be safely destroyed.

Explicit Console Group and Port Bindings Are Not Migrated

Bug ID 6781589: During a migration, any explicitly assigned console group and port are ignored, and a console with default properties is created for the target domain. This console is created using the target domain name as the console group and using any available port on the first virtual console concentrator (vcc) device in the control domain. If there is a conflict with the default group name, the migration fails.

Recovery: To restore the explicit console properties following a migration, unbind the target domain, and manually set the desired properties using the ldm set-vcons command.

VIO DR Operations Ignore the Force (-f) Option

Bug ID 6703127: Virtual input/output (VIO) dynamic reconfiguration (DR) operations ignore the -f (force) option in CLI commands.

ldmconfig Might Cause the Root File System of the Control Domain to Become Full and Halt the System

Bug ID 6848114: ldmconfig can run on a system that does not have file systems of sufficient capacity to contain the virtual disks for the created domains. In this situation, an error message is issued. However, ldmconfig permits you to continue to use the disks that are in /ldoms/disks to deploy the configuration. This situation could cause the root file system of the control domain to become full and halt the system.

Workaround: Do the following:

  1. Exit the Configuration Assistant by typing q or by typing Ctrl-C.

  2. Add more file systems of adequate capacity.

  3. Rerun the ldmconfig command.

Guest Domain Sometimes Fails to Make Proper Domain Services Connection to the Control Domain

Bug ID 6839787: Sometimes, a guest domain that runs at least the Oracle Solaris 10 10/08 OS does not make a proper Domain Services connection to a control domain that runs the Oracle Solaris 10 5/09 OS.

Domain Services connections enable features such as dynamic reconfiguration (DR), FMA, and power management (PM). Such a failure occurs when the guest domain is booted, so rebooting the domain usually clears the problem.

Workaround: Reboot the guest domain.

ldmd Might Dump Core If Multiple set-vcpu Operations Are Performed on the Control Domain While It Is in Delayed Reconfiguration Mode

Bug ID 6697096: Under certain circumstances, when multiple ldm set-vcpu operations are performed on the control domain while it is in delayed reconfiguration mode, ldmd might abort and be restarted by the Service Management Facility (SMF).

While the control domain is in delayed reconfiguration mode, take care when attempting an ldm set-vcpu operation. A single ldm set-vcpu operation will succeed, but a second ldm set-vcpu operation might cause the ldmd daemon to dump core.

Workaround: Reboot the control domain before you attempt the second ldm set-vcpu operation.

Virtual Network Devices Are Not Created Properly on the Control Domain

Bug ID 6836587: Sometimes ifconfig indicates that the device does not exist after you add a virtual network or virtual disk device to a domain. This situation might occur as the result of the /devices entry not being created.

Although this should not occur during normal operation, the error was seen when the instance number of a virtual network device did not match the instance number listed in /etc/path_to_inst file.

For example:

# ifconfig vnet0 plumb
ifconfig: plumb: vnet0: no such interface

The instance number of a virtual device is shown under the DEVICE column in the ldm list output:

# ldm list -o network primary
NAME             
primary          

MAC
    00:14:4f:86:6a:64

VSW
    NAME         MAC               NET-DEV DEVICE   DEFAULT-VLAN-ID PVID VID MTU  MODE  
    primary-vsw0 00:14:4f:f9:86:f3 nxge0   switch@0 1               1        1500        

NETWORK
    NAME   SERVICE              DEVICE    MAC               MODE PVID VID MTU  
    vnet1  primary-vsw0@primary network@0 00:14:4f:f8:76:6d      1        1500

The instance number (0 for both the vnet and vsw shown previously) can be compared with the instance number in the path_to_inst file to ensure that they match.

# egrep '(vnet|vsw)' /etc/path_to_inst
"/virtual-devices@100/channel-devices@200/virtual-network-switch@0" 0 "vsw"
"/virtual-devices@100/channel-devices@200/network@0" 0 "vnet"

Workaround: In the case of mismatching instance numbers, remove the virtual network or virtual switch device. Then, add them again by explicitly specifying the instance number required by setting the id property.

You can also manually edit the /etc/path_to_inst file. See the path_to_inst(4) man page.


Caution

Caution - Be aware of the warning contained in the man page that states “changes should not be made to /etc/path_to_inst without careful consideration.”


Cannot Connect to Migrated Domain's Console Unless vntsd Is Restarted

Bug ID 6757486: Occasionally, after a domain has been migrated, it is not possible to connect to the console for that domain.

Workaround: Restart the vntsd SMF service to enable connections to the console:

# svcadm restart vntsd

Note - This command will disconnect all active console connections.


I/O Domain or Guest Domain Panics When Booting From e1000g

Bug ID 6808832: You can configure a maximum of two domains with dedicated PCI-E root complexes on systems such as the Sun Fire T5240. These systems have two UltraSPARC T2+ CPUs and two I/O root complexes.

pci@500 and pci@400 are the two root complexes in the system. The primary domain will always contain at least one root complex. A second domain can be configured with an unassigned or unbound root complex.

The pci@400 fabric (or leaf) contains the onboard e1000g network card. The following circumstances could lead to a domain panic:

Avoid the following network devices if they are configured in a non-primary domain:

/pci@400/pci@0/pci@c/network@0,1
/pci@400/pci@0/pci@c/network@0

When these conditions are true, the domain will panic with a PCI-E Fatal error.

Avoid such a configuration, or if the configuration is used, do not boot from the listed devices.

Guest Domain Might Fail to Successfully Reboot When a System Is in Power Management Elastic Mode

Bug ID 6853273: While a system is in power management elastic mode, rebooting a guest domain might produce the following warning messages and fail to boot successfully:

WARNING: /virtual-devices@100/channel-devices@200/disk@0:
Sending packet to LDC, status: -1
WARNING: /virtual-devices@100/channel-devices@200/disk@0:
Can't send vdisk read request!
WARNING: /virtual-devices@100/channel-devices@200/disk@0:
Timeout receiving packet from LDC ... retrying

Workaround: If you see these warnings, perform one of the workarounds in the following order:

ldm Commands Are Slow to Respond When Several Domains Are Booting

Bug ID 6855079: An ldm command might be slow to respond when several domains are booting. If you issue an ldm command at this stage, the command might appear to hang. Note that the ldm command will return after performing the expected task. After the command returns, the system should respond normally to ldm commands.

Workaround: Avoid booting many domains simultaneously. However, if you must boot several domains at once, refrain from issuing further ldm commands until the system returns to normal. For instance, wait for about two minutes on Sun SPARC Enterprise T5140 and T5240 Servers and for about four minutes on the Sun SPARC Enterprise T5440 Server or Netra T5440 Server.

Spurious ds_ldc_cb: LDC READ event Message Seen When Rebooting the Control Domain or a Guest Domain

Bug ID 6846889: When rebooting the control domain or a guest domain, the following warning message might be logged on the control domain and on the guest domain that is rebooting:

WARNING: ds@0: ds_ldc_cb: LDC READ event while port not up

Workaround: You can ignore this message.

Migrated Domain With MAUs Contains Only One CPU When Target OS Does Not Support DR of Cryptographic Units

Bug ID 6904849: Starting with the Logical Domains 1.3 release, a domain can be migrated even if it has one or more cryptographic units bound to it.

In the following circumstances, the target machine will only contain one CPU after the migration is completed:

After the migration completes, the target domain will resume successfully and be operational, but will be in a degraded state (just one CPU).

Workaround: Prior to the migration, remove the cryptographic unit or units from the source machine that runs Logical Domains 1.3.

Mitigation: To avoid this issue, perform one or both of these steps:

Confusing Migration Failure Message for Real Address Memory Bind Failures

Bug ID 6904240: In certain situations, a migration fails with the following error message, and ldmd reports that it was not possible to bind the memory needed for the source domain. This situation can occur even if the total amount of available memory on the target machine is greater than the amount of memory being used by the source domain (as shown by ldm ls-devices -a mem).

Unable to bind 29952M memory region at real address 0x8000000
Domain Migration of LDom ldg0 failed

Cause: This failure is due the inability to meet congruence requirements between the Real Address (RA) and the Physical Address (PA) on the target machine.

Workaround: Stop the domain and perform the migration as a cold migration. You can also reduce the size of the memory on the guest domain by 128 Mbytes, which might permit the migration to proceed while the domain is running.

Newly Added NIU/XAUI Adapters Are Not Visible to Host OS If Logical Domains Is Configured

Bug ID 6829016: When Logical Domains is configured on a system and you add another XAUI network card, the card is not visible after the machine is powercycled.

Recovery: To make the newly added XAUI visible in the control domain, perform the following steps:

  1. Set and clear a dummy variable in the control domain.

    The following commands use a dummy variable called fix-xaui:

    # ldm set-var fix-xaui=yes primary
    # ldm rm-var fix-xaui primary
  2. Save the modified configuration to the SP, replacing the current configuration.

    The following commands use a configuration name of config1:

    # ldm rm-spconfig config1
    # ldm add-spconfig config1
  3. Perform a reconfiguration reboot of the control domain.

    # reboot -- -r

    At this time, you can configure the newly available network or networks for use by Logical Domains.

Dynamically Removing All the Cryptographic Units From a Domain Causes SSH to Terminate

Bug ID 6897743: If all the hardware cryptographic units are dynamically removed from a running domain, the cryptographic framework fails to seamlessly switch to the software cryptographic providers, and kills all the ssh connections.

Recovery: Re-establish the ssh connections after all the cryptographic units are removed from the domain.

Workaround: Set UseOpenSSLEngine=no in the /etc/ssh/sshd_config file on the server side, and run the svcadm restart ssh command.

Then, all ssh connections will no longer use the hardware cryptographic units (and thus not benefit from the associated performance improvements), and ssh connections would not be disconnected when the cryptographic units are removed.

Atlas PCI Express Dual 10-Gigabit Ethernet Fiber Card Shows Four Subdevices in ldm list-io -l Output

Bug ID 6892229: When you run the ldm ls-io -l command on a system that has an Atlas PCI Express Dual 10-Gigabit Ethernet Fiber card (X1027A-Z) installed, the output might show the following:

primary# ldm ls-io -l
...
pci@500/pci@0/pci@c PCIE5 OCC primary
network@0
network@0,1
ethernet
ethernet

The output shows four subdevices even though this Ethernet card has only two ports. This anomaly occurs because this card has four PCI functions. Two of these functions are disabled internally and appear as ethernet in the ldm ls-io -l output.

Workaround: You can ignore the ethernet entries in the ldm ls-io -l output.

ldm set-domain -i Command Fails to Completely Configure the Domain as Specified by the XML File

Bug ID 6965623: The ldm set-domain -i filename command cannot create a logical domain based on the configuration described by the specified XML file. The command parses the ldom_info node from the file, but fails to parse the other resource nodes, such as cpu, mau, and memory.

Workaround: Use the ldm add-domain -i filename command to add the domain and to configure the domain based on the specified XML file.

ldmp2v Command: ufsdump Archiving Method Is No Longer Used

Bug ID 6924550: When you use the ufsdump archiving method and a virtual disk backed by a file on a UFS file system, the system might hang when you run the ldmp2v prepare command. The ufsdump archiving method is no longer available for the ldmp2v collect command and has been superseded by the Flash archiving method. For compatibility with previously created ufsdump archives, the ldmp2v prepare command can still use existing ufsdump archives. However, it is better to create a new archive by using the Flash archiving method. This potential hang also applies when you manually restore ufsdump archives in preparation for the ldmp2v prepare -R /altroot command when the virtual disk is a file on a UFS file system.

ldmp2v convert Command: VxVM Warning Messages During Boot

Running Veritas Volume Manager (VxVM) 5.x on the Oracle Solaris 10 OS is the only supported (tested) version for the Oracle VM Server for SPARC P2V tool. Older versions of VxVM, such as 3.x and 4.x running on the Solaris 8 and Solaris 9 operating systems, might also work. In those cases, the first boot after running the ldmp2v convert command might show warning messages from the VxVM drivers. You can ignore these messages. You can remove the old VRTS* packages after the guest domain has booted.

Boot device: disk0:a File and args:
SunOS Release 5.10 Version Generic_139555-08 64-bit
Copyright 1983-2009 Sun Microsystems, Inc. All rights reserved.
Use is subject to license terms.
Hostname: normaal
Configuring devices.
/kernel/drv/sparcv9/vxdmp: undefined symbol ?romp?
WARNING: mod_load: cannot load module ?vxdmp?
WARNING: vxdmp: unable to resolve dependency, module ?misc/ted? not found
/kernel/drv/sparcv9/vxdmp: undefined symbol ?romp?
WARNING: mod_load: cannot load module ?vxdmp?
WARNING: vxdmp: unable to resolve dependency, module ?misc/ted? not found
/kernel/drv/sparcv9/vxio: undefined symbol ?romp?
WARNING: mod_load: cannot load module ?vxio?
WARNING: vxio: unable to resolve dependency, module ?drv/vxdmp? not found
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
WARNING: VxVM vxspec V-5-0-0 vxspec: vxio not loaded. Aborting vxspec load
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
WARNING: VxVM vxspec V-5-0-0 vxspec: vxio not loaded. Aborting vxspec load
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
WARNING: VxVM vxspec V-5-0-0 vxspec: vxio not loaded. Aborting vxspec load
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
WARNING: VxVM vxspec V-5-0-0 vxspec: vxio not loaded. Aborting vxspec load
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
WARNING: VxVM vxspec V-5-0-0 vxspec: vxio not loaded. Aborting vxspec load
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
WARNING: VxVM vxspec V-5-0-0 vxspec: vxio not loaded. Aborting vxspec load
WARNING: vxspec : CANNOT INITIALIZE vxio DRIVER
NOTICE: VxVM not started
Dynamic Reconfiguration of MTU Values of Virtual Network Devices Sometimes Fails

Bug ID 6936833: If you modify the maximum transmission unit (MTU) of a virtual network device on the control domain, a delayed reconfiguration operation is triggered. If you subsequently cancel the delayed reconfiguration, the MTU value for the device is not restored to the original value.

Recovery: Rerun the ldm set-vnet command to set the MTU to the original value. Resetting the MTU value puts the control domain into delayed reconfiguration mode, which you need to cancel. The resulting MTU value is now the original, correct MTU value.

# ldm set-vnet mtu=orig-value vnet1 primary
# ldm cancel-op reconf primary
Active Guest Domain Erroneously Shown to Be in Transition Mode

Bug ID 6961910: An active guest domain that runs the Oracle Solaris OS appears to be in the transition state rather than in the normal state in the following circumstances:

Workaround: Reboot the domain to return it to the normal state.

ldmconfig -c Command Fails Silently

Bug ID 6922142: When you run the ldmconfig -c command to check the Oracle Solaris OS media for valid packages, the check fails silently. The appropriate virtual devices to access the CD-ROM are not created, so the newly created logical domains cannot boot from or install from a CD-ROM device.

The ldmconfig command does not work if a user installs the Oracle Solaris OS from a CD.

Workaround: If you boot from a CD, you must manually configure the logical domains. Manually create the CD-ROM cdsdev and virtual disk. For example, for the Oracle Solaris 10/09 OS and the ldg0 domain, you would type the following commands:

# ldm add-vdsdev options=ro /vol/dev/dsk/c0t0d0/sol_10_1009_sparc cdrom0@primary-vds0
# ldm add-vdisk cdrom cdrom0@primary-vds0 ldg0
Memory DR Is Disabled Following a Canceled Migration

Bug ID 6956431: After an Oracle Solaris 10 9/10 domain has been suspended as part of a migration operation, memory dynamic reconfiguration (DR) is disabled. This applies not only when the migration is successful, but also when the migration has been canceled, despite the fact that the domain remains on the source machine.

Migration of a Guest Domain That Has Hybrid I/O-Enabled Virtual Network Devices Panics the Service Domain

Bug ID 6972633: The service domain panics when performing a warm migration of a guest domain. The source system in the migration is a SPARC T3-1 that has the NIU hybrid I/O capability.

The problem can occur when all of the following conditions are met:

A guest domain that has hybrid I/O enabled for a virtual network interface shows hybrid in the MODE column as follows:

# ldm list -o network ldg1
...
NAME    SERVICE             ID  DEVICE     MAC                MODE    PVID  MTU
vnet2    niu-vsw@primary     1  network@1  00:14:4f:fa:9e:89  hybrid  1    1500

However, the hybrid I/O resource is only assigned if the following command shows any output on the guest domain:

# kstat -p nxge

Workaround: Perform the following steps:

  1. Obtain the current configuration of the virtual network device.

    This step ensures that replumbing the interface is error-free.

    # ifconfig vnet1
  2. Unplumb the virtual network interface on the guest domain prior to the migration.

    # ifconfig vnet1 unplumb
  3. Perform the migration.

  4. Plumb the interface.

    # ifconfig vnet1 plumb
Migrated Guest Domain Hangs If It Has CPUs That Are Offline, Faulted, or Spare

Bug ID 6965789: This hang occurs on a guest domain under the following conditions:

Use the psrinfo command to determine whether any CPUs in the guest domain have been marked as offline, faulted, or spare. This command shows the state of CPUs.

Workaround: To avoid the problem on a guest domain that has offline or spare CPUs, use the psradm command to bring all CPUs online before attempting to migrate the domain.

ldmconfig Command Does Not Save Virtual Disk Service Devices for Bound Domains

Bug ID 6967799: When you use the ldmconfig command to create a service processor (SP) configuration, the default primary-with-clients configuration is created before the command adds the vol0 or vol1 virtual disk service device.

Migration of a Domain That Has an Enabled Default DRM Policy Results in a Target Domain Being Assigned All Available CPUs

Bug ID 6968507: Following the migration of an active domain, CPU utilization in the migrated domain can increase dramatically for a short period of time. If there is a dynamic resource managment (DRM) policy in effect for the domain at the time of the migration, the Logical Domains Manager might begin to add CPUs. In particular, if the vcpu-max and attack properties were not specified when the policy was added, the default value of unlimited causes all the unbound CPUs in the target machine to be added to the migrated domain.

Recovery: No recovery is necessary. After the CPU utilization drops below the upper limit that is specified by the DRM policy, the Logical Domains Manager automatically removes the CPUs.

Uncooperative Oracle Solaris Domain Migration Can Be Blocked If cpu0 Is Offline

Bug ID 6965758: The migration of an active domain can fail if it is running a release older than the Oracle Solaris 10 10/09 OS and the lowest numbered CPU in the domain is in the offline state. The operation fails when the Logical Domains Manager uses CPU DR to reduce the domain to a single CPU. In doing so, the Logical Domains Manager attempts to remove all but the lowest CPU in the domain, but as that CPU is offline, the operation fails.

Workaround: Before attempting the migration, ensure that the lowest numbered CPU in the domain is in the online state.

Sometimes DR Requests Fail to Remove All Requested CPUs

Bug ID 6493140: Sometimes, the Oracle Solaris OS is unable to use DR to remove all the requested CPUs. When this problem occurs, you see error messages similar to the following:

Removal of cpu 10 failed

Recovery: Issue a subsequent request to remove the number of CPUs that failed to be removed the first time. Such a retry generally succeeds.

When Re-creating a Guest Domain From an XML File, the Domain Should be Created by Using the CPU Allocation Unit and Not the Whole Core Number

Bug ID 6977065: A guest domain that is re-created by means of an XML file incorrectly uses the value of the max-cores property instead of the CPU allocation unit.

For example, a guest domain that is initially configured and bound with 10 whole cores (max-cores=10) is later reconfigured to use three whole cores. At this time, the guest domain's constraints are saved to an XML file. Later, the guest domain is unbound and destroyed, and then re-created by using the saved XML file. Instead of the guest domain being configured with three whole cores, it is configured with 10 whole cores based on the value of the max-cores property.

Inactive or Bound Domains Allow the Coexistence of the Whole-Core Constraint With an Enabled DRM Policy

Bug ID 6978606: The whole-core constraint and a DRM policy are not permitted to be enabled simultaneously. However, depending on the order in which you enable the constraint or policy, the ldm list command shows that both are enabled simultaneously.

If you enable the whole-core constraint on an inactive or bound domain that already has an active DRM policy, both appear as enabled in the ldm list output. In this situation, the DRM policy is disabled when you make the domain active.

ldm set-vcc port-range Command While in Delayed Reconfiguration Mode Fails to Hold the New port-range Value After Reboot

Bug ID 6975322: When the primary domain is in delayed reconfiguration mode, the ldm set-vcc port-range command does not work.

Workaround: Ensure that the primary domain is not in delayed reconfiguration mode before running the ldm set-vcc port-range command.

Using ldm stop -a Command on Domains in a Master-Slave Relationship Leaves the Slave With the stopping Flag Set

Bug ID 6979574: When a reset dependency is created, an ldm stop -a command might result in a domain with a reset dependency being restarted instead of only stopped.

Workaround: First, issue the ldm stop command to the master domain. Then, issue the ldm stop command to the slave domain. If the initial stop of the slave domain results in a failure, issue the ldm stop -f command to the slave domain.

Guest Domain Might Panic After the Domain Is Started With the ldm start Command

Bug ID 6979007: If you attempt to start a guest domain in elastic mode with the ldm start command, the guest domain might panic during boot.

This panic is rare and can only be seen from the guest domain's console. The control domain is not affected, and there is no indication that the guest domain panicked on the control domain console.

Workaround: No action is necessary. The guest domain resets and reboots successfully.

ldmp2v convert Command: The drd Daemon Is Not Always Enabled

Bug ID 6924931: When you convert a sun4u system that runs the Oracle Solaris 10 OS to a logical domain, the drd daemon is not enabled automatically after conversion. As a result, DR is not possible for the converted domain.

Workaround: Manually enable the drd daemon after the conversion by running the following command:

# svcadm enable svc:/platform/sun4v/drd:default
ldc_close: (0xb) unregister failed, 11 Warning Messages

Bug ID 6610702: You might see the following warning message on the system console or in the system log:

ldc_close: (0xb) unregister failed, 11

Note that the number in parentheses is the Oracle Solaris internal channel number, which might be different for each warning message.

Workaround: You can ignore these messages.

Memory DR Removal Operations With Unplumbed NIU nxge Instances Can Cause a Kernel Panic, Panic - Generated at user request

Bug ID 6977504: When a memory dynamic reconfiguration (DR) removal command (ldm rm-mem or ldm set-mem) is issued that affects a domain that owns an NIU device, the domain might panic if the NIU's nxge driver instance is not plumbed. The panic might look similar to the following:

panic[cpu0]/thread=2a10001fca0: Panic - Generated at user request

Recovery: Reboot the affected domain.

Workaround: Before performing any memory DR operations on a domain that owns an NIU device, configure the domain to plumb all NIU devices at boot time, and then reboot. NIU devices should be plumbed at boot time and remain plumbed before you attempt any memory DR removal operations.

To determine if NIU devices are present on a domain, check the /etc/path_to_inst file for references to NIU devices.

# grep nxge /etc/path_to_inst
"/niu@80/network@0" 0 "nxge"
"/niu@80/network@1" 1 "nxge"

The /niu@80 string in the two lines indicates that nxge instances 0 and 1 are for NIU devices. The nxge0 and nxge1 devices should be plumbed at boot time. These devices must remain in the plumbed state when you perform any memory DR operations on the domain by using ldm rm-mem or ldm set-mem command.

To plumb the nxge0 and nxge1 instances without configuring network parameters for the devices, create the /etc/hostname.nxge0 and /etc/hostname.nxge1 files as follows:

# touch /etc/hostname.nxge0
# touch /etc/hostname.nxge1
Memory DR Removal Operations With Multiple Plumbed NIU nxge Instances Can Hang Indefinitely and Never Complete

Bug ID 6983279: When multiple NIU nxge instances are plumbed on a domain, the ldm rm-mem and ldm set-mem commands, which are used to remove memory from the domain, might never complete. To determine whether the problem has occurred during a memory removal operation, monitor the progress of the operation with the ldm list -o status command. You might have encountered this problem if the progress percentage remains constant for several minutes.

Recovery: Cancel the ldm rm-mem or ldm set-mem command.

Workaround: Cancel the ldm rm-mem or ldm set-mem command, and check if a sufficient amount of memory was removed. If not, a subsequent memory removal command to remove a smaller amount of memory might complete successfully.

If the problem has occurred on the primary domain, do the following:

  1. Start a delayed reconfiguration operation on the primary domain.

    # ldm start-reconf primary
  2. Assign the desired amount of memory to the domain.

  3. Reboot the primary domain.

If the problem occurred on another domain, stop the domain before adjusting the amount of memory that is assigned to the domain.

Direct I/O Configuration Is Not Properly Rebuilt by the ldm init-system Command

Bug ID 6980444: The existing I/O devices on the primary domain are not properly removed when the primary domain is reconfigured by using the ldm init-system command with an XML file. This command might cause binding failures on the guest domain. These failures occur if the primary domain is still binding PCIe leaf node devices to the primary domain.

Workaround: Use the ldm rm-io command to manually remove the PCIe leaf node devices from the primary domain. Then, reboot the primary domain before you bind and start the guest domains.

Guest Domains Cannot Boot From an Exported DVD Device

Bug ID 6981081: When a bootable physical CD or DVD is exported as a virtual disk, the virtual CD or DVD might not be bootable from the guest domain that uses it. Also, the boot might fail with an error similar to the following:

{0} ok boot /virtual-devices@100/channel-devices@200/disk@1:f
Boot device: /virtual-devices@100/channel-devices@200/disk@1:f  File and args:
Bad magic number in disk label
ERROR: /virtual-devices@100/channel-devices@200/disk@1: Can't open disk label package
ERROR: boot-read fail
Can't open boot device

Whether this problem occurs depends on the type of physical CD or DVD drive that is installed on the system.

Emulex HBA Permanently Destroyed If the Firmware Upgrade Is Interrupted

Bug ID 6950462: The Emulex driver automatically performs a firmware upgrade of the Emulex host bus adapter (HBA) when the driver and HBA firmware levels do not match. This might occur when the Emulex HBA is assigned to an I/O domain by means of the direct I/O feature. If the firmware upgrade of an Emulex HBA is interrupted, the card is permanently destroyed. So, do not assign any Emulex HBA to an I/O domain with the direct I/O feature.

Block of Dynamically Added Memory Can Only Be Dynamically Removed as a Whole

A block of dynamically added memory is subject to a restriction in which that memory block can only be dynamically removed as a whole. That is, a subset of that memory block cannot be dynamically removed.

This situation could occur if a domain with a small memory size is dynamically grown to a much larger size, as the following example shows:

# ldm list ldom1
NAME  STATE FLAGS   CONS VCPU MEMORY UTIL UPTIME
ldom1 active -n---- 5000 2    1G     0.4% 23h

# ldm add-mem 16G ldom1

# ldm rm-mem 8G ldom1
Memory removal failed because all of the memory is in use.

# ldm rm-mem 16G ldom1

# ldm list ldom1
NAME  STATE FLAGS   CONS VCPU MEMORY UTIL UPTIME
ldom1 active -n---- 5000 2    1G     0.4% 23h

Workaround: Dynamically add memory in smaller amounts to reduce the probability that this condition will occur.

Recovery: Reboot the domain.

Cryptographic Units Cannot Be Removed From the primary Domain

Bug ID 6978843: Sometimes, when you attempt to dynamically remove cryptographic units, the following message is issued:

# ldm set-crypto 0 primary
Aug 20 13:02:27 guest1 ncp: WARNING: ncp0: ncp_mau_unconfig:
unable to find MAU for cpu 112
Aug 20 13:02:27 guest1 ncp: WARNING: ncp0: ncp_mau_unconfig:
unable to find MAU for cpu 104

Workaround: Determine whether any CPUs are faulted, and if they are, mark them as being online.

# psrinfo
# psradm -n 0-127

Use delayed reconfiguration to remove the cryptographic units.

# ldm start-reconf primary
# ldm set-crypto 0 primary
# reboot
SPARC T3-1: Detect And Handle Disks That Are Accessible Through Multiple Direct I/O Paths

Bug ID 6984008: A SPARC T3-1 system can be installed with dual-ported disks, which can be accessed by two different direct I/O devices. In this case, assigning these two direct I/O devices to different domains can cause the disks to be used by both domains and impact each other based on the actual usage of those disks.

Workaround: Do not assign direct I/O devices that have access to the same set of disks to different I/O domains. The steps to determine if you have dual-ported disks on T3-1 system are as follows:

Determine whether the system has dual-ported disks by running the following command on the SP:

-> show /SYS/SASBP

If the output includes the following fru_description value, the corresponding system has dual-ported disks:

fru_description = BD,SAS2,16DSK,LOUISE

When dual disks are found to be present in the system, ensure that both of the following direct I/O devices are always assigned to the same domain:

pci@400/pci@1/pci@0/pci@4  /SYS/MB/SASHBA0
pci@400/pci@2/pci@0/pci@4  /SYS/MB/SASHBA1
System Powers Off When a Memory Request Results in a Hypervisor Abort

Bug ID 6986076: A hypervisor abort could cause your system to power off if all of the following conditions are met:

Workaround: Do one of the following if you plan to issue one or more memory requests to the control domain while it is in delayed reconfiguration mode:

nxge Driver Supports Only Eight Hybrid I/O Shares in a Guest Domain

Bug ID 6805126: Each guest domain that uses hybrid I/O cannot support the maximum number of 32 hybrid I/O shares. If you attempt to configure more than 8 virtual network devices in hybrid mode on a guest domain, the system might panic.

Workaround: Configure no more than eight virtual network devices in hybrid mode on a single guest domain.

Documentation Errata

This section contains documentation errors that have been found too late to resolve for the Oracle VM Server for SPARC 2.0 release.

Incorrect Parameter Names in the Input/Output Bus Table

Bug ID 6843196: “Input/Output Bus Table (IOBusTable)” on page 31 of the Logical Domains (LDoms) MIB 1.0.1 Administration Guide shows incorrect parameter names.

IOBusDevName should be IOBusName, and IOBusDevPath should be IOBusPath.

virtinfo(1M) and Related API Man Pages Missing From the Oracle Solaris 10 9/10 Release

The virtinfo(1M), v12n(3EXT), and libv12n(3LIB) man pages are missing from the Oracle Solaris 10 9/10 man pages.

For the Oracle VM Server for SPARC 2.0 release, see these man pages in the Oracle VM Server for SPARC 2.0 Reference Manual.

Cannot Upgrade From Logical Domains 1.0 to Oracle VM Server for SPARC 2.0

Upgrading to Oracle VM Server for SPARC 2.0 Software in Oracle VM Server for SPARC 2.0 Administration Guide refers to an upgrade path where you can upgrade an UltraSPARC T1 system that is running the Logical Domains 1.0 software to the Oracle VM Server for SPARC 2.0 software.

You cannot perform this kind of upgrade because the UltraSPARC T1 platform is not supported by the Oracle VM Server for SPARC 2.0 software.