On 20/07/17 02:53, Sage Weil wrote:
On Wed, 19 Jul 2017, Mark Kirkwood wrote:
One (I think) new thing compared to the 12.1.0 is that restarting the
services
blitzes the modified crushmap, and we get back to:
$ sudo ceph osd tree
ID CLASS WEIGHT TYPE NAME UP/DOWN REWEIGHT PRI-AFF
-1 0.32996 root default
-2 0.08199 host ceph1
0 hdd 0.02399 osd.0 up 1.00000 1.00000
4 hdd 0.05699 osd.4 up 1.00000 1.00000
-3 0.08299 host ceph2
1 hdd 0.02399 osd.1 up 1.00000 1.00000
5 hdd 0.05899 osd.5 up 1.00000 1.00000
-4 0.08199 host ceph3
2 hdd 0.02399 osd.2 up 1.00000 1.00000
6 hdd 0.05699 osd.6 up 1.00000 1.00000
-5 0.08299 host ceph4
3 hdd 0.02399 osd.3 up 1.00000 1.00000
7 hdd 0.05899 osd.7 up 1.00000 1.00000
...and all the PG are remapped again. Now I might have just missed this
happening with 12.1.0 - but I'm (moderately) confident that I did
restart
stuff and not see this happening. For now I've added:
osd crush update on start = false
to my ceph.conf to avoid being caught by this.
Actually setting the above does *not* prevent the crushmap getting changed.
Can you share teh output of 'ceph osd metadata 0' vs 'cpeh osd metadata
4'? I'm not sure why it's getting the class wrong. I haven't seen this
on my cluster (it's bluestore; maybe that's the difference).
Yes, and it is quite interesting: osd 0 is filestore on hdd, osd 4 is
bluestore on ssd but (see below) the metadata suggests ceph thinks it is hdd
(the fact that the hosts are VMs might be not helping here):
$ sudo ceph osd metadata 0
{
"id": 0,
"arch": "x86_64",
"back_addr": "192.168.122.21:6806/1712",
"backend_filestore_dev_node": "unknown",
"backend_filestore_partition_path": "unknown",
"ceph_version": "ceph version 12.1.1
(f3e663a190bf2ed12c7e3cda288b9a159572c800) luminous (rc)",
"cpu": "QEMU Virtual CPU version 1.7.0",
"distro": "ubuntu",
"distro_description": "Ubuntu 16.04.2 LTS",
"distro_version": "16.04",
"filestore_backend": "xfs",
"filestore_f_type": "0x58465342",
"front_addr": "192.168.122.21:6805/1712",
"hb_back_addr": "192.168.122.21:6807/1712",
"hb_front_addr": "192.168.122.21:6808/1712",
"hostname": "ceph1",
"kernel_description": "#106-Ubuntu SMP Mon Jun 26 17:54:43 UTC 2017",
"kernel_version": "4.4.0-83-generic",
"mem_swap_kb": "1047548",
"mem_total_kb": "2048188",
"os": "Linux",
"osd_data": "/var/lib/ceph/osd/ceph-0",
"osd_journal": "/var/lib/ceph/osd/ceph-0/journal",
"osd_objectstore": "filestore",
"rotational": "1"
}
$ sudo ceph osd metadata 4
{
"id": 4,
"arch": "x86_64",
"back_addr": "192.168.122.21:6802/1488",
"bluefs": "1",
"bluefs_db_access_mode": "blk",
"bluefs_db_block_size": "4096",
"bluefs_db_dev": "253:32",
"bluefs_db_dev_node": "vdc",
"bluefs_db_driver": "KernelDevice",
"bluefs_db_model": "",
"bluefs_db_partition_path": "/dev/vdc2",
"bluefs_db_rotational": "1",
"bluefs_db_size": "63244840960",
"bluefs_db_type": "hdd",
"bluefs_single_shared_device": "1",
"bluestore_bdev_access_mode": "blk",
"bluestore_bdev_block_size": "4096",
"bluestore_bdev_dev": "253:32",
"bluestore_bdev_dev_node": "vdc",
"bluestore_bdev_driver": "KernelDevice",
"bluestore_bdev_model": "",
"bluestore_bdev_partition_path": "/dev/vdc2",
"bluestore_bdev_rotational": "1",
"bluestore_bdev_size": "63244840960",
"bluestore_bdev_type": "hdd",
"ceph_version": "ceph version 12.1.1
(f3e663a190bf2ed12c7e3cda288b9a159572c800) luminous (rc)",
"cpu": "QEMU Virtual CPU version 1.7.0",
"distro": "ubuntu",
"distro_description": "Ubuntu 16.04.2 LTS",
"distro_version": "16.04",
"front_addr": "192.168.122.21:6801/1488",
"hb_back_addr": "192.168.122.21:6803/1488",
"hb_front_addr": "192.168.122.21:6804/1488",
"hostname": "ceph1",
"kernel_description": "#106-Ubuntu SMP Mon Jun 26 17:54:43 UTC 2017",
"kernel_version": "4.4.0-83-generic",
"mem_swap_kb": "1047548",
"mem_total_kb": "2048188",
"os": "Linux",
"osd_data": "/var/lib/ceph/osd/ceph-4",
"osd_journal": "/var/lib/ceph/osd/ceph-4/journal",
"osd_objectstore": "bluestore",
"rotational": "1"
}