BS2020 LXC to LXD Notes
When I set up BS2020 a year ago I was new to LXC and LXD (both of which are used at present because I wanted separate disk pools and network for infrastructure and development(/deployment). I believe that once we move from 16.04 to 18.04 we should be able to remove LXC from the equation. Regardless the initial pools were set up using lxd init as described in the install notes for BS2020.
root@bs2020:~# lxd init
Name of the storage backend to use (dir or zfs) [default=zfs]:
Create a new ZFS pool (yes/no) [default=yes]? yes
Name of the new ZFS pool [default=lxd]: lxd4infra
Would you like to use an existing block device (yes/no) [default=no]? yes
Path to the existing block device: /dev/sde1
Would you like LXD to be available over the network (yes/no) [default=no]?
Do you want to configure the LXD bridge (yes/no) [default=yes]? no
root@bs2020:~# lxd init
... create new zfs pool and use all of /dev/sdd1 do not configure bridge ...
root@bs2020:~# dpkg-reconfigure -p medium lxd
... no yes br1 ... use existing bridge...
root@bs2020:~#
When we moved disks from the original server to the new one the OS renumerated the disks so that the boot disk is at /dev/sdc1 (bay 2?) and our archive disk is in the last bay (bay 5). The remaining disks make up two zfs pools. In hindsight I would have preferred to initialize the original disks as entire disks as apposed to the first slice.
root@bs2020:~# df -k
...
/dev/sdc1 41921808 2140004 37629256 6% /
...
/dev/sdf1 480589544 1524692 454629192 1% /archive
...
root@bs2020:~#
root@bs2020:~# zpool status
pool: lxd4dev
state: ONLINE
scan: scrub repaired 0 in 0h6m with 0 errors on Sun May 13 00:30:55 2018
config:
NAME STATE READ WRITE CKSUM
lxd4dev ONLINE 0 0 0
sdd1 ONLINE 0 0 0
sdb ONLINE 0 0 0
sde ONLINE 0 0 0
errors: No known data errors
pool: lxd4infra
state: ONLINE
scan: scrub repaired 0 in 0h2m with 0 errors on Sun May 13 00:26:10 2018
config:
NAME STATE READ WRITE CKSUM
lxd4infra ONLINE 0 0 0
sda1 ONLINE 0 0 0
errors: No known data errors
Disks sdb and sde were added to the dev/deployment pool as follows.
zpool add -f lxd4dev /dev/sde
zpool add -f lxd4dev /dev/sdb
As the disks were previously used they should have been wiped first.
wipefs -a /dev/sdf
New Disks New Errors
Since adding the two new disks I keep getting io errors on one of them. They do not seem to be causing any data errors however.
ZFS has detected an io error:
eid: 106
class: io
host: bs2020
time: 2018-05-31 13:45:36-0700
vtype: disk
vpath: /dev/sdb1
vguid: 0x04009EB732FC8852
cksum: 0
read: 0
write: 0
pool: lxd4dev
Backing up containers using zfs
Link dump
- http://www.digithink.com/serverdocs/FunWithLinuxDisks
- https://www.thegeekdiary.com/zfs-tutorials-creating-zfs-snapshot-and-clones/