Migrate FileServer to Debian #44

Closed
opened 2020-11-13 10:55:51 +01:00 by moepman · 9 comments
Owner

The FileServer (bowle) should be migrated to Debian as soon as OpenZFS 2.0 is released (this should allow for an import of the existing FreeBSD ZFS pool).

The FileServer (bowle) should be migrated to Debian as soon as OpenZFS 2.0 is released (this should allow for an import of the existing FreeBSD ZFS pool).
moepman added the
enhancement
label 2020-11-13 10:55:51 +01:00
moepman added a new dependency 2020-11-13 10:56:57 +01:00
Author
Owner

This should probably wait for Debian bullseye (to be released this year) since that will natively support OpenZFS 2.0

This should probably wait for Debian bullseye (to be released this year) since that will natively support OpenZFS 2.0
moepman added this to the (deleted) milestone 2021-05-06 22:01:27 +02:00
Author
Owner

I guess this needs to maybe happen before the official Release of Debian bullseye. Today the USB-Stick containing FreeBSD 12 has died.

I guess this needs to maybe happen before the official Release of Debian bullseye. Today the USB-Stick containing FreeBSD 12 has died.
moepman modified the milestone from (deleted) to 2021 Q2 2021-06-07 13:05:09 +02:00
Author
Owner

A very basic role (with still a few TODOs left) is in place as of b2442be2d8

A very basic role (with still a few TODOs left) is in place as of b2442be2d85ba66e4a9811f34f0f4280d2f7d7b2
Author
Owner

Current status: ZFS and NFS installed, with a basic NFS config in place. NFS exports with backup space for the Proxmox-based VM-Hosts seem to be working.

However the zpool did not recognize one of the disks:

root@bowle ~ # zpool status
  pool: pool
 state: DEGRADED
status: One or more devices could not be used because the label is missing or
        invalid.  Sufficient replicas exist for the pool to continue
        functioning in a degraded state.
action: Replace the device using 'zpool replace'.
   see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-4J
  scan: resilvered 726G in 02:12:48 with 0 errors on Sat Dec 26 17:52:41 2020
config:

        NAME                     STATE     READ WRITE CKSUM
        pool                     DEGRADED     0     0     0
          raidz2-0               DEGRADED     0     0     0
            sda                  ONLINE       0     0     0
            sdb                  ONLINE       0     0     0
            sdd                  ONLINE       0     0     0
            1267470862749452811  UNAVAIL      0     0     0  was /dev/ada3
            sdh                  ONLINE       0     0     0
            sde                  ONLINE       0     0     0
            sdf                  ONLINE       0     0     0
            sdg                  ONLINE       0     0     0

errors: No known data errors

/dev/sdc is there but it can't be set online or used with the zpool replace command - maybe its headers need to be wiped and it needs to be re-added as a blank disk?

Current status: ZFS and NFS installed, with a basic NFS config in place. NFS exports with backup space for the Proxmox-based VM-Hosts seem to be working. However the zpool did not recognize one of the disks: ``` root@bowle ~ # zpool status pool: pool state: DEGRADED status: One or more devices could not be used because the label is missing or invalid. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Replace the device using 'zpool replace'. see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-4J scan: resilvered 726G in 02:12:48 with 0 errors on Sat Dec 26 17:52:41 2020 config: NAME STATE READ WRITE CKSUM pool DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 sda ONLINE 0 0 0 sdb ONLINE 0 0 0 sdd ONLINE 0 0 0 1267470862749452811 UNAVAIL 0 0 0 was /dev/ada3 sdh ONLINE 0 0 0 sde ONLINE 0 0 0 sdf ONLINE 0 0 0 sdg ONLINE 0 0 0 errors: No known data errors ``` `/dev/sdc` is there but it can't be set online or used with the `zpool replace` command - maybe its headers need to be wiped and it needs to be re-added as a blank disk?
Author
Owner

Update:

root@bowle ~ # fdisk /dev/sdb
[...]

Command (m for help): g

Created a new GPT disklabel (GUID: 325080D4-1682-3144-83F1-C22C16A8C79D).
The device contains 'ddf_raid_member' signature and it will be removed by a write command. See fdisk(8) man page and --wipe option for more details.

Command (m for help): w

The partition table has been altered.
Calling ioctl() to re-read partition table.
Syncing disks.

After this and a reboot (without it would still assume the disk being in use) later I could finally replace the disk with the invalid label:

root@bowle ~ # zpool replace pool 1267470862749452811 /dev/sdc
root@bowle ~ # zpool status
  pool: pool
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
        continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
  scan: resilver in progress since Mon Jun  7 13:54:19 2021
        792G scanned at 4.92G/s, 105G issued at 669M/s, 5.80T total
        12.9G resilvered, 1.77% done, 02:28:49 to go
config:

        NAME                       STATE     READ WRITE CKSUM
        pool                       DEGRADED     0     0     0
          raidz2-0                 DEGRADED     0     0     0
            sdb                    ONLINE       0     0     0
            sda                    ONLINE       0     0     0
            sdd                    ONLINE       0     0     0
            replacing-3            DEGRADED     0     0     0
              1267470862749452811  UNAVAIL      0     0     0  was /dev/ada3
              sdc                  ONLINE       0     0     0  (resilvering)
            sdh                    ONLINE       0     0     0
            sde                    ONLINE       0     0     0
            sdf                    ONLINE       0     0     0
            sdg                    ONLINE       0     0     0

errors: No known data errors
Update: ``` root@bowle ~ # fdisk /dev/sdb [...] Command (m for help): g Created a new GPT disklabel (GUID: 325080D4-1682-3144-83F1-C22C16A8C79D). The device contains 'ddf_raid_member' signature and it will be removed by a write command. See fdisk(8) man page and --wipe option for more details. Command (m for help): w The partition table has been altered. Calling ioctl() to re-read partition table. Syncing disks. ``` After this and a reboot (without it would still assume the disk being in use) later I could finally replace the disk with the invalid label: ``` root@bowle ~ # zpool replace pool 1267470862749452811 /dev/sdc root@bowle ~ # zpool status pool: pool state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Mon Jun 7 13:54:19 2021 792G scanned at 4.92G/s, 105G issued at 669M/s, 5.80T total 12.9G resilvered, 1.77% done, 02:28:49 to go config: NAME STATE READ WRITE CKSUM pool DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 sdb ONLINE 0 0 0 sda ONLINE 0 0 0 sdd ONLINE 0 0 0 replacing-3 DEGRADED 0 0 0 1267470862749452811 UNAVAIL 0 0 0 was /dev/ada3 sdc ONLINE 0 0 0 (resilvering) sdh ONLINE 0 0 0 sde ONLINE 0 0 0 sdf ONLINE 0 0 0 sdg ONLINE 0 0 0 errors: No known data errors ```
Author
Owner

Looks good, running a scrub just to be sure.

Looks good, running a scrub just to be sure.
Author
Owner

All good.

root@bowle ~ # zpool status
  pool: pool
 state: ONLINE
  scan: scrub repaired 0B in 01:38:57 with 0 errors on Mon Jun  7 18:51:01 2021
config:

        NAME        STATE     READ WRITE CKSUM
        pool        ONLINE       0     0     0
          raidz2-0  ONLINE       0     0     0
            sdb     ONLINE       0     0     0
            sda     ONLINE       0     0     0
            sdd     ONLINE       0     0     0
            sdc     ONLINE       0     0     0
            sdh     ONLINE       0     0     0
            sde     ONLINE       0     0     0
            sdf     ONLINE       0     0     0
            sdg     ONLINE       0     0     0

errors: No known data errors
All good. ``` root@bowle ~ # zpool status pool: pool state: ONLINE scan: scrub repaired 0B in 01:38:57 with 0 errors on Mon Jun 7 18:51:01 2021 config: NAME STATE READ WRITE CKSUM pool ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 sdb ONLINE 0 0 0 sda ONLINE 0 0 0 sdd ONLINE 0 0 0 sdc ONLINE 0 0 0 sdh ONLINE 0 0 0 sde ONLINE 0 0 0 sdf ONLINE 0 0 0 sdg ONLINE 0 0 0 errors: No known data errors ```
Author
Owner

The only things left wrt to the hardware is to buy an USB3 to SATA adapter and replace the loaned SSD with the donated one from exxess.

The ansible Role itself is tracked at #12

The only things left wrt to the hardware is to buy an USB3 to SATA adapter and replace the loaned SSD with the donated one from exxess. The ansible Role itself is tracked at #12
Author
Owner

SSD and USB3-SATA adapter have been swapped yesterday.

SSD and USB3-SATA adapter have been swapped yesterday.
Sign in to join this conversation.
No Milestone
No Assignees
1 Participants
Notifications
Due Date
The due date is invalid or out of range. Please use the format 'yyyy-mm-dd'.

No due date set.

Blocks
Reference: infra/ansible#44
No description provided.