This procedure deploys the FLAR on the target system.
The target guest domain name is solaris10 in this example.
FLAR (SourceGlobal.flar)
Oracle Solaris 10 ISO image (sol-10-u11-ga-sparc-dvd.iso)
root@SourceGlobal# cd /ovas1 root@SourceGlobal# ls -rtlh *.gz *.flar -rw-r--r-- 1 root root 11G Jul 30 07:58 dbzone.gz -rw-r--r-- 1 root root 4.8G Jul 30 08:08 dbzone_db_binary.gz -rw-r--r-- 1 root root 5.8G Jul 30 09:43 asm1.img.gz -rw-r--r-- 1 root root 5.8G Jul 30 09:45 asm2.img.gz -rw-r--r-- 1 root root 59M Jul 30 11:09 redo.ufsdump.gz -rw-r--r-- 1 root root 185M Jul 30 11:14 archive.ufsdump.gz -rw-r--r-- 1 root root 32G Jul 30 13:14 SourceGlobal.flar
Refer to instructions in Lift the Source Environment to the Shared Storage.
The ldmp2vz_convert command restores the source OS ZFS root file system to the target guest domain in a new boot environment. This restores the zone configuration and updates the new boot environment with SVR4 packages that are required for the sun4v target system.
Syntax:
ldmp2vz_convert -f FLAR_filename -i iso_filename -b boot_environment
Where:
-f FLAR_filename – Is the file name of the FLAR created in Lift the Source Environment to the Shared Storage.
-i iso_filename – Is the file name of the Oracle Solaris ISO file.
-b boot_environment – Is the name of the guest domain boot environment.
Example:
root@TargetGuestDom# cd /ovas1
root@TargetGuestDom# ldmp2vz_convert -f SourceGlobal.flar -i ./Downloads/sol-10-u11-ga-sparc-dvd.iso -b mybe
Renaming /var/yp to /var/yp.pre_p2v
Running: lucreate -s - -n mybe -p rpool -o /var/log/ldmp2vz/1669/lucreate.log
See /var/log/ldmp2vz/1669/lucreate.log for detailed output
Determining types of file systems supported
Validating file system requests
Preparing logical storage devices
Preparing physical storage devices
Configuring physical storage devices
Configuring logical storage devices
INFORMATION: No BEs are configured on this system.
The device </dev/dsk/c0d0s0> is not a root device for any boot environment; cannot get BE ID.
Running: zpool set autoexpand=on rpool
Renaming zfs dataset rpool/export to rpool/export_pre_p2v
Running: luupgrade -f -n mybe -s /tmp/p2v.sCa4qd/iso_unbundled -a /ovas1/SourceGlobal.flar -o /var/log/ldmp2vz/1669/luupgrade1.log
See /var/log/ldmp2vz/1669/luupgrade1.log for detailed output
67352 blocks
cannot open 'dbzone_db_binary': dataset does not exist
cannot open 'dbzone_db_binary': dataset does not exist
/a
INFO: Removing sun4u packages:
SUNWcakr
SUNWcar
SUNWced
SUNWcpc
SUNWcpr
SUNWdscpr
SUNWdscpu
SUNWefc
SUNWifb
SUNWjfb
SUNWkvm
SUNWm64
SUNWnxge
SUNWpfb
SUNWpstl
SUNWsckmu
SUNWstc
Removal of <SUNWcakr> was successful.
Removal of <SUNWcar> was successful.
Removal of <SUNWced> was successful.
Removal of <SUNWcpc> was successful.
Removal of <SUNWcpr> was successful.
Removal of <SUNWdscpr> was successful.
Removal of <SUNWdscpu> was successful.
Removal of <SUNWefc> was successful.
/a/dev/fb*: No such file or directory
Removal of <SUNWifb> was successful.
/a/dev/fb*: No such file or directory
Removal of <SUNWjfb> was successful.
Removal of <SUNWkvm> was successful.
Removal of <SUNWm64> was successful.
Removal of <SUNWnxge> was successful.
Removal of <SUNWpfb> was successful.
Removal of <SUNWpstl> was successful.
Removal of <SUNWsckmu> was successful.
Removal of <SUNWstc> was successful.
Running: luupgrade -u -n mybe -s /tmp/p2v.sCa4qd/iso_unbundled -k /tmp/no-autoreg -o /var/log/ldmp2vz/1669/luupgrade2.log
See /var/log/ldmp2vz/1669/luupgrade2.log for detailed output
67352 blocks
cannot open 'dbzone_db_binary': dataset does not exist
/a
Extracting configuration files
x etc/defaultdomain, 11 bytes, 1 tape blocks
x etc/nsswitch.conf, 434 bytes, 1 tape blocks
x etc/pam.conf, 3454 bytes, 7 tape blocks
x var/ldap, 0 bytes, 0 tape blocks
x var/ldap/secmod.db, 131072 bytes, 256 tape blocks
x var/ldap/Oracle_SSL_CA_G2.pem, 1830 bytes, 4 tape blocks
x var/ldap/VTN-PublicPrimary-G5.pem, 1732 bytes, 4 tape blocks
x var/ldap/cachemgr.log, 10274 bytes, 21 tape blocks
x var/ldap/key3.db, 131072 bytes, 256 tape blocks
x var/ldap/oracle_ssl_cert.pem, 2134 bytes, 5 tape blocks
x var/ldap/ldap_client_cred, 204 bytes, 1 tape blocks
x var/ldap/ldap_client_file, 2935 bytes, 6 tape blocks
x var/ldap/cert8.db, 65536 bytes, 128 tape blocks
Disabling Oracle Configuration Manager service in /a/var/svc/manifest/application/management/ocm.xml
Running: luumount mybe
Running: luactivate mybe
A Live Upgrade Sync operation will be performed on startup of boot environment <mybe>.
**********************************************************************
The target boot environment has been activated. It will be used when you
reboot. NOTE: You MUST NOT USE the reboot, halt, or uadmin commands. You
MUST USE either the init or the shutdown command when you reboot. If you
do not use either init or shutdown, the system will not boot using the
target BE.
**********************************************************************
In case of a failure while booting to the target BE, the following process
needs to be followed to fallback to the currently working boot environment:
1. Enter the PROM monitor (ok prompt).
2. Boot the machine to Single User mode using a different boot device
(like the Solaris Install CD or Network). Examples:
At the PROM monitor (ok prompt):
For boot to Solaris CD: boot cdrom -s
For boot to network: boot net -s
3. Mount the Current boot environment root slice to some directory (like
/mnt). You can use the following commands in sequence to mount the BE:
zpool import rpool
zfs inherit -r mountpoint rpool/ROOT/s10s_u11wos_24a
zfs set mountpoint=<mountpointName> rpool/ROOT/s10s_u11wos_24a
zfs mount rpool/ROOT/s10s_u11wos_24a
4. Run <luactivate> utility with out any arguments from the Parent boot
environment root slice, as shown below:
<mountpointName>/sbin/luactivate
5. luactivate, activates the previous working boot environment and
indicates the result.
6. umount /mnt
7. zfs set mountpoint=/ rpool/ROOT/s10s_u11wos_24a
8. Exit Single User mode and reboot the machine.
**********************************************************************
Modifying boot archive service
Activation of boot environment <mybe> successful.
Run:
init 6
which will cause a reboot.
After reboot run:
svcadm disable ocm
svcadm enable ldap/client
Then install any desired patch(es). Follow instructions to reboot again
if necessary. After rebooting, wait until the system boots to
multi-user-server milestone.
Check this using the command: svcs multi-user-server
and look for state 'online'.
Then copy and restore any required additional filesystems, for example non-ZFS filesystems, from system SourceGlobal.
root@TargetGuestDom# shutdown -i 6
root@TargetGuestDom# svcadm disable ocm root@TargetGuestDom# svcadm enable ldap/client
root@TargetGuestDom# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / native shared - webzone configured /rpool/webzone native shared - dbzone configured /zones/dbzone native shared
root@TargetGuestDom# mount -F nfs TargetControlDom:/ovas1 /ovas1