+1-323-909-4740 blog@hegars.com

Found a set of disk in my pile that came out of a Buffalo NAS

root@debian:~# mdadm --detail /dev/md127
/dev/md127:
           Version : 1.2
     Creation Time : Thu May 28 11:42:20 2015
        Raid Level : raid5
        Array Size : 1953259520 (1862.77 GiB 2000.14 GB)
     Used Dev Size : 976629760 (931.39 GiB 1000.07 GB)
      Raid Devices : 3
     Total Devices : 3
       Persistence : Superblock is persistent

     Intent Bitmap : Internal

       Update Time : Thu Oct  3 01:21:52 2019
             State : active 
    Active Devices : 3
   Working Devices : 3
    Failed Devices : 0
     Spare Devices : 0

            Layout : left-symmetric
        Chunk Size : 512K

Consistency Policy : bitmap

              Name : buffalo:0
              UUID : 817c66c9:ab5aba3d:46cf3be0:c46fd224
            Events : 11591

    Number   Major   Minor   RaidDevice State
       0       8       17        0      active sync   /dev/sdb1
       1       8       49        1      active sync   /dev/sdd1
       2       8       33        2      active sync   /dev/sdc1
root@debian:~# pvdisplay /dev/md127          
  --- Physical volume ---
  PV Name               /dev/md127
  VG Name               raid5
  PV Size               <1.82 TiB / not usable 4.00 MiB
  Allocatable           yes 
  PE Size               4.00 MiB
  Total PE              476869
  Free PE               338357
  Allocated PE          138512
  PV UUID               2UPreL-vdQI-NdXn-xfhR-BxgN-FADg-UpYjVn
root@debian:~# vgdisplay raid5               
  --- Volume group ---
  VG Name               raid5
  System ID             
  Format                lvm2
  Metadata Areas        1
  Metadata Sequence No  9
  VG Access             read/write
  VG Status             resizable
  MAX LV                0
  Cur LV                7
  Open LV               5
  Max PV                0
  Cur PV                1
  Act PV                1
  VG Size               <1.82 TiB
  PE Size               4.00 MiB
  Total PE              476869
  Alloc PE / Size       138512 / 541.06 GiB
  Free  PE / Size       338357 / 1.29 TiB
  VG UUID               ADkPZw-R27s-6FUZ-3LW7-JUOV-oPyW-55pWu4
root@debian:~# lvdisplay raid5               
  --- Logical volume ---
  LV Path                /dev/raid5/root
  LV Name                root
  VG Name                raid5
  LV UUID                CRKUp4-a5lj-QeI2-mdiX-ifxY-3OH5-wVWGOw
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-05-28 11:50:46 +1000
  LV Status              available
  # open                 1
  LV Size                37.25 GiB
  Current LE             9536
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:2
   
  --- Logical volume ---
  LV Path                /dev/raid5/swap
  LV Name                swap
  VG Name                raid5
  LV UUID                iGclU8-Fm3T-GSgD-L8AP-UJDk-Y5H3-iiQmWo
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-05-28 11:51:00 +1000
  LV Status              available
  # open                 0
  LV Size                3.81 GiB
  Current LE             976
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:3
   
  --- Logical volume ---
  LV Path                /dev/raid5/zpool1
  LV Name                zpool1
  VG Name                raid5
  LV UUID                clahFF-RXYF-QsQ3-xBgG-rMa9-y0Ac-8zDdit
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-05-28 13:36:22 +1000
  LV Status              available
  # open                 1
  LV Size                100.00 GiB
  Current LE             25600
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:4
   
  --- Logical volume ---
  LV Path                /dev/raid5/zpool2
  LV Name                zpool2
  VG Name                raid5
  LV UUID                3v95G3-Uev9-V3LI-PyYt-WzXu-kdch-fNMsjP
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-05-28 13:36:26 +1000
  LV Status              available
  # open                 1
  LV Size                100.00 GiB
  Current LE             25600
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:5
   
  --- Logical volume ---
  LV Path                /dev/raid5/zpool3
  LV Name                zpool3
  VG Name                raid5
  LV UUID                y4McZn-hLW3-jQR2-Pzkz-9QEc-Jd5p-TZ9idI
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-05-28 13:36:29 +1000
  LV Status              available
  # open                 1
  LV Size                100.00 GiB
  Current LE             25600
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:6
   
  --- Logical volume ---
  LV Path                /dev/raid5/zpool4
  LV Name                zpool4
  VG Name                raid5
  LV UUID                DOO9th-5B32-9kX6-NkwG-4Dbx-AKcN-83XiAT
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-05-28 13:36:32 +1000
  LV Status              available
  # open                 1
  LV Size                100.00 GiB
  Current LE             25600
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:7
   
  --- Logical volume ---
  LV Path                /dev/raid5/macmini
  LV Name                macmini
  VG Name                raid5
  LV UUID                XL6d9V-SGfH-6m16-WH8E-JobI-oGn9-HuGDjL
  LV Write Access        read/write
  LV Creation host, time buffalo, 2015-06-01 13:46:25 +1000
  LV Status              available
  # open                 0
  LV Size                100.00 GiB
  Current LE             25600
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     4096
  Block device           253:8

Upgrading ZFS/zPool

Disks are now in a Debian 10 machine and the original version was v4 ZFS and v23 zPool created via ” zfs-fuse 0.7.2 “

ZFS

root@debian:~# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
tank  57.3G   328G  57.3G  /tank
root@debian:~# zfs get version
NAME  PROPERTY  VALUE    SOURCE
tank  version   4        -
root@debian:/# zfs upgrade 
This system is currently running ZFS filesystem version 5.

The following filesystems are out of date, and can be upgraded.  After being
upgraded, these filesystems (and any 'zfs send' streams generated from
subsequent snapshots) will no longer be accessible by older software versions.


VER  FILESYSTEM
---  ------------
 4   tank
root@debian:/# zfs upgrade tank
1 filesystems upgraded

zPool

root@debian:~# zpool get version
NAME  PROPERTY  VALUE    SOURCE
tank  version   23       local
root@debian:~# zpool list tank 
NAME   SIZE  ALLOC   FREE  EXPANDSZ   FRAG    CAP  DEDUP  HEALTH  ALTROOT
tank   398G  73.2G   325G         -      -    18%  1.00x  ONLINE  -
root@debian:~# zpool status
  pool: tank
 state: ONLINE
status: The pool is formatted using a legacy on-disk format.  The pool can
        still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
        pool will no longer be accessible on software that does not support
        feature flags.
  scan: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        tank        ONLINE       0     0     0
          zpool1    ONLINE       0     0     0
          zpool2    ONLINE       0     0     0
          zpool3    ONLINE       0     0     0
          zpool4    ONLINE       0     0     0

errors: No known data errors
root@debian:/# zpool upgrade tank 
This system supports ZFS pool feature flags.

Successfully upgraded 'tank' from version 23 to feature flags.
Enabled the following features on 'tank':
  async_destroy
  empty_bpobj
  lz4_compress
  multi_vdev_crash_dump
  spacemap_histogram
  enabled_txg
  hole_birth
  extensible_dataset
  embedded_data
  bookmarks
  filesystem_limits
  large_blocks
  large_dnode
  sha512
  skein
  edonr
  userobj_accounting
root@debian:/# zpool get version  
NAME  PROPERTY  VALUE    SOURCE
tank  version   -        default
root@debian:/# zpool status
  pool: tank
 state: ONLINE
  scan: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        tank        ONLINE       0     0     0
          zpool1    ONLINE       0     0     0
          zpool2    ONLINE       0     0     0
          zpool3    ONLINE       0     0     0
          zpool4    ONLINE       0     0     0

errors: No known data errors