[OpenIndiana-discuss] wiki.openindiana.org is unavailable?? Temporary or permanent, gone, ceased??

Aurélien Larcher aurelien.larcher at gmail.com
Tue Sep 20 13:54:52 UTC 2016


But I would personally use the root split method over the old one.

On Tue, Sep 20, 2016 at 3:10 PM, Мартин Бохниг via openindiana-discuss
<openindiana-discuss at openindiana.org> wrote:
> Hello,
>
> I have no login credentials nor anything else other than ML access (and never needed it nor asked for).
> But here is some info you may find useful:
>
> A) As always in all periods/epochs/ages, I took wget copies of all important sites.
> So if anybody ever needs any mailing list copy, bugs database or download central stuff from opensolaris.org, schillix, belenix, opensparc.net,  or oi or illumos etc, I can help you in most cases
>
> B) It took me 10 to 15 seconds from your question to google to finding the links you are referring to (down atm) to google cache.
> The content a few lines lower  ...
>
> C) While not most but definitely at lot of most lightweight stuff should always make it into webarchive.org.
>
>
> Now your requested content:
>
>
> MfG, %martin
>
>
> How to migrate the root pool
> Skip to end of metadata
> *  Page restrictions apply
> *  Added by  Gary Mills , last edited by  Predrag Zečević on Dec 05, 2013   ( view change ) Go to start of metadata
> *
> I recently wanted to migrate the root pool to a new device.  This turned out to be easy to do, using extisting facilities.  The original root pool was on an old 80-gig disk.  This system also had a data pool on a newer 1 TB disk.  Here's what the `format' command showed for them:
>        0. c2t0d0 <Unknown-Unknown-0001 cyl 9726 alt 2 hd 255 sec 63>
>           /pci at 0,0/pci1043,8389 at 11/disk at 0,0
>        1. c2t2d0 <ATA-ST31000524AS-JC4B-931.51GB>
>           /pci at 0,0/pci1043,8389 at 11/disk at 2,0I
> *  wanted to migrate the root pool to a new SSD.  The `format' command was available to prepare the SSD.  I could use the `zpool' command to create the pool on that new device, and `beadm' and `installgrub' to perform the migration.  That part worked out nicely.  I had to use a variety of commands to complete the migration.
> *  Add the SSD: Just shut down the computer, install the SSD hardware, and boot the system.  Here's the new output from `format':
>        0. c2t0d0 <Unknown-Unknown-0001 cyl 9726 alt 2 hd 255 sec 63>
>           /pci at 0,0/pci1043,8389 at 11/disk at 0,0
>        1. c2t1d0 <ATA-SanDiskSDSSDP06-0 cyl 9966 alt 2 hd 224 sec 56>
>           /pci at 0,0/pci1043,8389 at 11/disk at 1,0
>        2. c2t2d0 <ATA-ST31000524AS-JC4B-931.51GB>
>           /pci at 0,0/pci1043,8389 at 11/disk at 2,0
> *  Prepare the SSD: Create the fdisk partition within `format':
>     format> fdisk
>     No fdisk table exists. The default partition for the disk is:
>       a 100% "SOLARIS System" partition
>     Type "y" to accept the default partition,  otherwise type "n" to edit the
>      partition table.
>     y
>
> *  Create the slice:
>     partition> 0
>     Part      Tag    Flag     Cylinders        Size            Blocks
>       0 unassigned    wm       0               0         (0/0/0)            0
>     Enter partition id tag[unassigned]: root
>     Enter partition permission flags[wm]:
>     Enter new starting cyl[1]: 3
>     Enter partition size[0b, 0c, 3e, 0.00mb, 0.00gb]: $
>     partition> p
>     Current partition table (unnamed):
>     Total disk cylinders available: 9965 + 2 (reserved cylinders)
>     Part      Tag    Flag     Cylinders        Size            Blocks
>       0       root    wm       3 - 9964       59.59GB    (9962/0/0) 124963328
>       1 unassigned    wm       0               0         (0/0/0)            0
>       2     backup    wu       0 - 9964       59.61GB    (9965/0/0) 125000960
>       3 unassigned    wm       0               0         (0/0/0)            0
>       4 unassigned    wm       0               0         (0/0/0)            0
>       5 unassigned    wm       0               0         (0/0/0)            0
>       6 unassigned    wm       0               0         (0/0/0)            0
>       7 unassigned    wm       0               0         (0/0/0)            0
>       8       boot    wu       0 -    0        6.12MB    (1/0/0)        12544
>       9 unassigned    wm       0               0         (0/0/0)            0
>     partition> l
>     Ready to label disk, continue? y
> *  Get the root pool version:
>     # zpool get all rpool
>     NAME   PROPERTY               VALUE                  SOURCE
>     rpool  size                   74G                    -
>     ...
>     rpool  version                28                     local
>
> *  Try to create the new root pool, with a new pool name:
>     # zpool create -o version=28 rpool1 c2t1d0s0
>     invalid vdev specification
>     use '-f' to override the following errors:
>     /dev/dsk/c2t1d0s0 overlaps with /dev/dsk/c2t1d0s2
>
> *  Try again with the force option:
>     # zpool create -f -o version=28 rpool1 c2t1d0s0
>     # zpool list
>     NAME     SIZE  ALLOC   FREE  EXPANDSZ    CAP  DEDUP  HEALTH  ALTROOT
>     dpool    928G  85.6G   842G     2.50M     9%  1.00x  ONLINE  -
>     rpool     74G  7.05G  66.9G         -     9%  1.00x  ONLINE  -
>     rpool1  59.5G   108K  59.5G         -     0%  1.00x  ONLINE  -
>
> *  Create the BE, on the new device with a new name:
>     # beadm create -p rpool1 oi_151a6x
>     WARNING: menu.lst file /rpool1/boot/grub/menu.lst does not exist,
>              generating a new menu.lst file
>     Created successfully
>
> *  Verify that it exists:
>     # beadm list
>     BE          Active Mountpoint Space Policy Created
>     oi_151a6    NR     /          5.98G static 2012-09-13 16:33
>     oi_151a6x   R      -          4.15G static 2013-06-06 15:55
>     openindiana -      -          13.5M static 2012-09-13 08:55
>
> *  Install the boot blocks:
>     # installgrub -m /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c2t1d0s0
>     Updating master boot sector destroys existing boot managers (if any).
>     continue (y/n)?y
>     stage2 written to partition 0, 277 sectors starting at 50 (abs 12594)
>     stage1 written to partition 0 sector 0 (abs 12544)
>     stage1 written to master boot sector
>
> *  Change the BIOS boot order by shutting the system down and entering the BIOS setup.  Then put the SSD first in the boot order and reboot.
> *  At this point, I upgraded to oi_151a7.  This confirmed that the new root pool was functional.  Here's the initial boot environment:
>     # beadm list
>     BE          Active Mountpoint Space Policy Created
>     oi_151a6    R      -          6.01G static 2012-09-13 16:33
>     oi_151a6x   NR     /          4.33G static 2013-06-06 15:55
>     openindiana -      -          13.5M static 2012-09-13 08:55
>
> *  Upgrade:
>     # pkg image-update --be-name oi_151a7
>     WARNING: The boot environment being modified is not the active one.  Changes    made in the active BE will not be reflected on the next boot.
>                 Packages to update: 895
>            Create boot environment: Yes
>     Create backup boot environment:  No
>     ...
>     A clone of oi_151a6x exists and has been updated and activated.
>     On the next boot the Boot Environment oi_151a7 will be
>     mounted on '/'.  Reboot when ready to switch to this updated BE.
>
> *  Check the BEs again:
>     # beadm list
>     BE          Active Mountpoint Space Policy Created
>     oi_151a6    R      -          6.01G static 2012-09-13 16:33
>     oi_151a6x   N      /          101K  static 2013-06-06 15:55
>     oi_151a7    R      -          5.31G static 2013-06-06 16:56
>     openindiana -      -          13.5M static 2012-09-13 08:55
>
> *  Shut down OS:
>     # init 5
>     updating //platform/i86pc/boot_archive
>     updating //platform/i86pc/amd64/boot_archive
>
> *  Press the `Power' button to reboot.  Confirm that the upgrade was successful.  Notice that there are still two active boot environments:
>     $ beadm list
>     BE          Active Mountpoint Space Policy Created
>     oi_151a6    R      -          6.01G static 2012-09-13 16:33
>     oi_151a6x   -      -          16.8M static 2013-06-06 15:55
>     oi_151a7    NR     /          5.33G static 2013-06-06 16:56
>     openindiana -      -          13.5M static 2012-09-13 08:55
>
> *  Some of the old root pool is still in use.  My home directory was on rpool/export/home/mills .  To simplify this migration, I decided to move it to the data pool.  First, create new filesystems on the data pool:
>     # zfs create dpool/export
>     # zfs create dpool/export/home
>
> *  My home directory in the /etc/passwd file was automounted to /home/mills from /export/home/mills .  The first thing I did was to copy it to /dpool/export/home/mills using `cpio'.  Then I edited /etc/passwd to change my home directory to /dpool/export/home/mills .  After that change, it was no longer automounted.  After a reboot, I confirmed that the old root pool was no longer needed for my home directory:
>     # zfs unmount rpool/export/home/mills
>     # zfs unmount rpool/export/home
>     # zfs unmount rpool/export
>
> *  Still, there are a few pieces left:
>     # zfs list | egrep 'dump|swap'
>     rpool/dump                895M  65.5G   895M  -
>     rpool/swap                952M  65.8G   637M  -
>
> *  To move the dump device, first get the properties of the old one:
>     $ zfs get all rpool/dump | egrep 'SOURCE|local'
>     NAME        PROPERTY                  VALUE                       SOURCE
>     rpool/dump  volsize                   895M                        local
>     rpool/dump  checksum                  off                         local
>     rpool/dump  compression               off                         local
>     rpool/dump  refreservation            none                        local
>     rpool/dump  dedup                     off                         local
>
> *  Create another one on rpool1:
>     # zfs create -o checksum=off -o compression=off -o refreservation=none -o dedup=off -V 895M rpool1/dump
>
> *  Try to move it:
>     # dumpadm -d /dev/zvol/dsk/rpool1/dump
>     dumpadm: dump device /dev/zvol/dsk/rpool1/dump is too small to hold a system dump
>     dump size 1812297728 bytes, device size 938475520 bytes
>     # dumpadm
>          Dump content: kernel pages
>            Dump device: /dev/zvol/dsk/rpool/dump (dedicated)
>     Savecore directory: /var/crash/ati
>       Savecore enabled: no
>        Save compressed: on
>
> *  Expand the volume and try again:
>     # dumpadm -d /dev/zvol/dsk/rpool1/dump
>           Dump content: kernel pages
>            Dump device: /dev/zvol/dsk/rpool1/dump (dedicated)
>     Savecore directory: /var/crash/ati
>       Savecore enabled: no
>        Save compressed: on
> *  Now, get the properties of the old swap device:
>     $ zfs get all rpool/swap | egrep 'SOURCE|local'
>     NAME        PROPERTY                  VALUE                       SOURCE
>     rpool/swap  volsize                   895M                        local
>     rpool/swap  refreservation            952M                        local
>
> *  Create a new one on rpool1:
>     # zfs create -o refreservation=952M -V 895M rpool1/swap
>
> *  Move the swap device by editing /etc/vfstab:
>     o Move the swap device by editing /etc/vfstab:
>     root at ati:/etc# cp -p vfstab vfstab-
>     root at ati:/etc# ex vfstab
>     root at ati:/etc# diff vfstab- vfstab
>     12c12
>     < /dev/zvol/dsk/rpool/swap  -               -               swap    -       no      -
>     ---
>     > /dev/zvol/dsk/rpool1/swap -               -               swap    -       no      -
>
> *  Reboot and confirm that rpool is no longer used:
>     # dumpadm
>           Dump content: kernel pages
>            Dump device: /dev/zvol/dsk/rpool1/dump (dedicated)
>     Savecore directory: /var/crash/ati
>       Savecore enabled: no
>        Save compressed: on
>     # swap -l
>     swapfile             dev    swaplo   blocks     free
>     /dev/zvol/dsk/rpool1/swap 96,2         8  1832952  1832952
>     # beadm list                                                          BE
>       Active Mountpoint Space Policy Created
>     oi_151a6    R      -          6.01G static 2012-09-13 16:33
>     oi_151a6x   -      -          16.8M static 2013-06-06 15:55
>     oi_151a7    NR     /          5.34G static 2013-06-06 16:56
>     openindiana -      -          13.5M static 2012-09-13 08:55
>     # zpool list
>     NAME     SIZE  ALLOC   FREE  EXPANDSZ    CAP  DEDUP  HEALTH  ALTROOT
>     dpool    928G  85.6G   842G     2.50M     9%  1.00x  ONLINE  -
>     rpool     74G  6.19G  67.8G         -     8%  1.00x  ONLINE  -
>     rpool1  59.5G  7.17G  52.3G         -    12%  1.00x  ONLINE  -
>
> *  Export the pool and observe the result:
>     # zpool export rpool
>     # zpool list
>     NAME     SIZE  ALLOC   FREE  EXPANDSZ    CAP  DEDUP  HEALTH  ALTROOT
>     dpool    928G  85.6G   842G     2.50M     9%  1.00x  ONLINE  -
>     rpool1  59.5G  7.18G  52.3G         -    12%  1.00x  ONLINE  -
>     # zfs list
>     NAME                    USED  AVAIL  REFER  MOUNTPOINT
>     dpool                  85.6G   828G    24K  /dpool
>     dpool/export           83.8G   828G    22K  /dpool/export
>     dpool/export/home      83.8G   828G  83.8G  /dpool/export/home
>     dpool/opt              1.82G   828G  1.82G  /dpool/opt
>     dpool/opt/local          21K   828G    21K  /dpool/opt/local
>     rpool1                 8.10G  50.5G  36.5K  /rpool1
>     rpool1/ROOT            5.17G  50.5G    31K  legacy
>     rpool1/ROOT/oi_151a6x  16.8M  50.5G  4.33G  /
>     rpool1/ROOT/oi_151a7   5.16G  50.5G  4.27G  /
>     rpool1/dump            2.00G  50.5G  2.00G  -
>     rpool1/swap             952M  51.4G    16K  -
>     # getent passwd mills
>     mills:x:107:10:Gary Mills:/dpool/export/home/mills:/bin/ksh
>     # beadm list
>     BE        Active Mountpoint Space Policy Created
>     oi_151a6x -      -          16.8M static 2013-06-06 15:55
>     oi_151a7  NR     /          5.34G static 2013-06-06 16:56
>
> *  I could have resumed automounting my home directory by changing the mount point of dpool/export to /export, but I decided to leave it the way it was.
> *  Here's another upgrade, just to confirm that the new root pool was correct:
>     # pkg image-update --be-name oi_151a8
>                 Packages to remove:  16
>                Packages to install:   6
>                 Packages to update: 879
>            Create boot environment: Yes
>     Create backup boot environment:  No
>     DOWNLOAD                                  PKGS       FILES    XFER (MB)
>     Completed                              901/901 22745/22745  566.2/566.2
>     PHASE                                        ACTIONS
>     Removal Phase                            13844/13844
>     Install Phase                            12382/12382
>     Update Phase                             23637/23637
>     PHASE                                          ITEMS
>     Package State Update Phase                 1780/1780
>     Package Cache Update Phase                   895/895
>     Image State Update Phase                         2/2
>     ...
>     root at ati:~# beadm list
>     BE        Active Mountpoint Space Policy Created
>     oi_151a6x -      -          16.8M static 2013-06-06 15:55
>     oi_151a7  N      /          11.4M static 2013-06-06 16:56
>     oi_151a8  R      -          8.76G static 2013-08-11 16:12
>     # bootadm list-menu
>     the location for the active GRUB menu is: /rpool1/boot/grub/menu.lst
>     default 2
>     timeout 30
>     0 oi_151a6x
>     1 oi_151a7
>     2 oi_151a8
>     # init 5
>
> *  Press the power switch to reboot.  The upgrade was successful, completing the migration to a new device.
>
> Labels:
> None  Edit Labels
> 3 Comments
> Hide/Show Comments
> *
> Dec 06, 2013
> Predrag Zečević
> Hi,
> I have also wanted to try SSD (Samsung SSD 840, 120 GB). My current rpool was on 160 GB HD 7200RPM. I have used slightly different approach, which worked (I am now writing this from system booted from SSD).
> First, I have created same partition layout as existing rpool had (slices 0, 2 and 8 - similar like in this example). BTW, I have attached SSD disk via USB docking station...
> Then I have created new pool (I have found disk ID using  format and  fdisk utilities in steps mentioned at the beginning of this page):
>
> Next phase is to take recursive snapshot of rpool and to send (verbose= -v , recursive= -R ) it to  and receive it (keep structure= -d , force= -F ) to new rpool (I have named it RPOOL):
>
> BTW, my installation has user home directories on second HD, as well as  /opt   directory. Boot disk ( rpool in such environment) has occupied 26 GB of space and system took 28 minutes under normal activities to send/receive pool...
> Now, we need to make new disk bootable. Check (compare and set  bootfs property of new root pool):
>
> After this, new pool has to be exported and grub installed:
>
> Now, you can shutdown system and shuffle disks. If you have put SSD disk to same controller, nothing to do... But, if you have changed location of it, then you have to fix BIOS boot order.
> I found easy enough to boot system FIRST from latest /hipster USB text installation image (less than 1GG, easy to create - and my installation IS /hipster one) in order to import copy of rpool under new name:
>
> After you have started reboot, skip step of booting from removable device and your system should be started from SSD now.
> My impression is that all this is NOT enough to have all benefits of SSD disk usage...
> Actually, I could not say that systems is significantly faster than boot from normal HD, but it might be needed to do some optimizations.
> This is how I did moved rpool to SSD (pardon me on my mistakes in English).
> Regards.
> P.S. Resources used (beside this page):
> *  http://ptribble.blogspot.de/2012/09/recursive-zfs-send-and-receive.html
> *  http://waddles.org/content/replicating-zfs-root-disks
>
> *  Permalink
> *
> Nov 15, 2014
> Jon Strabala
> Predrag
> You might be able to do this via "zpool split" without using snapshots ( I have not tried all these spteps ... yet )
> Lets's assume
> *  you have a rpool that is a bare drive or a mirrored set with the drive or one of the members "c1t0d0s0"
> *  you want to migrate the root pool to a new disk (same size or maybe bigger) to a new disk "c1t2d0s0"
> *  Note I'm not sure about any issues that might be caused via a 512 byte vs 4K disk sector mismatch
> so lets assume the sector sizes match on all the disks (old and new) .
> Note "zpool split" is not documented in in the illumos man page Bug #2897
> Step 1 - I imaging  a "cleaner procedure' with out relying on snapshots might be doing something like the following:
>   # zpool attach rpool c1t0d0s0 c1t2d0s0
> # zpool status rpool
> *** wait for resilver to complete ****
> Step 2 - Now split off the new device it's a perfect clone (by default it takes the last one added - but we could specify  c1t2d0s0 as the last arg )
> # zpool split rpool rpool2
>   # installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c1t2d0s0
> [Optional] You have your clone albiet with a different pool name, however what if your new drive is larger - your not using the space - but you can
> #  zfs list
> # zpool set autoexpand=on rpool2
> [Optional] not done yet look at how much space you can expand and use and then use your new space
> #  zpool get expandsize  rpool2
> #  zpool online -e  rpool2
> #  zfs list
> # zpool set autoexpand=off rpool2
> [Optional] at this point the new cloned disk may be bigger than the disks you cloned from if so no more using those old disks later as part of a mirror with the new disk
> Step 3. Time to set up the cloned disk to boot (we need to change it's pool name ) so shut down and power off
> # init 0
> Step 4. Remove the old OS drive (or drives) which is either the original stand alone disk or the entire original mirror set .
> Step 5. Boot from the  latest /hipster USB text image - only way I know of to change the pool name back to 'rpool'
> Step 6. Now import the device and change its name from rpool2 to rpool
> # zpool import -f rpool2 rpool
> # init 6
> IMHO Step 1 & 2 make a perfect clone except for the pool name - it would be cool if there was a zpool command to rename the split e.g. rpool2 to rpool WITHOUT bringing it online as it would have a "name" conflict and then you remove it offsite as a hot spare OS clone backup without rebooting to a /hipster image to rename it.
>
> *  Permalink
> *
> Sep 03, 2014
> Stefan Müller-Wilken
> Procedure for SPARC (as reported by igork on #oi-dev): need install zfs boot block by: installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c2t1d0s0
> *  Permalink
>
> Powered by a free  Atlassian Confluence Community License granted to OpenIndiana.  Evaluate Confluence today .
> *  Powered by  Atlassian Confluence 4.1.7 , the  Enterprise Wiki
> *  Printed by Atlassian Confluence 4.1.7, the Enterprise Wiki.
> *   ·   Report a bug
> *   ·   Atlassian News
>
>
>
>
>
>
>>Вторник, 20 сентября 2016, 10:25 UTC от "Hans J. Albertsson" <hans.j.albertsson at gmail.com>:
>>
>>Was going to refer to an old document on migrating the root pool, but I
>>get 503 Service unavailable from anywhere on wiki.openindiana.org.
>>
>>Is anyone  looking after this site? Will it reappear??
>>Is Gary Mills' short piece on migrating the root pool available elsewhere??
>>
> _______________________________________________
> openindiana-discuss mailing list
> openindiana-discuss at openindiana.org
> https://openindiana.org/mailman/listinfo/openindiana-discuss



-- 
---
Praise the Caffeine embeddings



More information about the openindiana-discuss mailing list