Image format 1 is still supported -- just trying to slowly move users off of it and onto image format 2 through lots of log message nagging. -- Jason Dillaman ----- Original Message ----- > From: "Stefan Lissmats" <stefan@xxxxxxxxxx> > To: "German Anders" <ganders@xxxxxxxxxxxx> > Cc: "ceph-users" <ceph-users@xxxxxxxxxxxxxx> > Sent: Tuesday, March 29, 2016 4:55:02 PM > Subject: Re: Scrubbing a lot > Ok, i also got the warning but was able to use it anyway. Could be blocked in > the new release of Jewel. Probably the more correct answer is to use the > other answer (to use - - image-features layering) but haven't tried that > myself. > Skickat från min Samsung-enhet > -------- Originalmeddelande -------- > Från: German Anders <ganders@xxxxxxxxxxxx> > Datum: 2016-03-29 22:48 (GMT+01:00) > Till: Stefan Lissmats <stefan@xxxxxxxxxx> > Kopia: Samuel Just <sjust@xxxxxxxxxx>, ceph-users <ceph-users@xxxxxxxxxxxxxx> > Rubrik: Re: Scrubbing a lot > mmmm it seems that the image-format option is deprecated: > # rbd --id cinder --cluster cephIB create e60host01v2 --size 100G > --image-format 1 --pool cinder-volumes -k > /etc/ceph/cephIB.client.cinder.keyring > rbd: image format 1 is deprecated > # rbd --cluster cephIB info e60host01v2 --pool cinder-volumes > 2016-03-29 16:45:39.073198 7fb859eb7700 -1 librbd::image::OpenRequest: RBD > image format 1 is deprecated. Please copy this image to image format 2. > rbd image 'e60host01v2': > size 102400 MB in 25600 objects > order 22 (4096 kB objects) > block_name_prefix: rb.0.37d7.238e1f29 > format: 1 > and the map operations still doesn't work :( > # rbd --cluster cephIB map e60host01v2 --pool cinder-volumes -k > /etc/ceph/cephIB.client.cinder.keyring > rbd: sysfs write failed > rbd: map failed: (5) Input/output error > also, I'm running kernel 3.19.0-39-generic > German > 2016-03-29 17:40 GMT-03:00 Stefan Lissmats < stefan@xxxxxxxxxx > : > > I agrree. I ran in to the same issue and the error massage is not that > > clear. > > Mapping with the kernel rbd client (rbd map) needs a quite new kernel to > > handle the new image format. The work-around is to use - - image-format 1 > > when creating the image. > > > -------- Originalmeddelande -------- > > > Från: Samuel Just < sjust@xxxxxxxxxx > > > > Datum: 2016-03-29 22:24 (GMT+01:00) > > > Till: German Anders < ganders@xxxxxxxxxxxx > > > > Kopia: ceph-users < ceph-users@xxxxxxxxxxxxxx > > > > Rubrik: Re: Scrubbing a lot > > > Sounds like a version/compatibility thing. Are your rbd clients really old? > > > -Sam > > > On Tue, Mar 29, 2016 at 1:19 PM, German Anders < ganders@xxxxxxxxxxxx > > > wrote: > > > > I've just upgrade to jewel, and the scrubbing seems to been corrected... > > > but > > > > now I'm not able to map an rbd on a host (before I was able to), > > > basically > > > > I'm getting this error msg: > > > > > > > > rbd: sysfs write failed > > > > rbd: map failed: (5) Input/output error > > > > > > > > # rbd --cluster cephIB create host01 --size 102400 --pool cinder-volumes > > > -k > > > > /etc/ceph/cephIB.client.cinder.keyring > > > > # rbd --cluster cephIB map host01 --pool cinder-volumes -k > > > > /etc/ceph/cephIB.client.cinder.keyring > > > > rbd: sysfs write failed > > > > rbd: map failed: (5) Input/output error > > > > > > > > Any ideas? on the /etc/ceph directory on the host I've: > > > > > > > > -rw-r--r-- 1 ceph ceph 92 Nov 17 15:45 rbdmap > > > > -rw-r--r-- 1 ceph ceph 170 Dec 15 14:47 secret.xml > > > > -rw-r--r-- 1 ceph ceph 37 Dec 15 15:12 virsh-secret > > > > -rw-r--r-- 1 ceph ceph 0 Dec 15 15:12 virsh-secret-set > > > > -rw-r--r-- 1 ceph ceph 37 Dec 21 14:53 virsh-secretIB > > > > -rw-r--r-- 1 ceph ceph 0 Dec 21 14:53 virsh-secret-setIB > > > > -rw-r--r-- 1 ceph ceph 173 Dec 22 13:34 secretIB.xml > > > > -rw-r--r-- 1 ceph ceph 619 Dec 22 13:38 ceph.conf > > > > -rw-r--r-- 1 ceph ceph 72 Dec 23 09:51 ceph.client.cinder.keyring > > > > -rw-r--r-- 1 ceph ceph 63 Mar 28 09:03 cephIB.client.cinder.keyring > > > > -rw-r--r-- 1 ceph ceph 526 Mar 28 12:06 cephIB.conf > > > > -rw------- 1 ceph ceph 63 Mar 29 16:11 cephIB.client.admin.keyring > > > > > > > > Thanks in advance, > > > > > > > > Best, > > > > > > > > German > > > > > > > > 2016-03-29 14:45 GMT-03:00 German Anders < ganders@xxxxxxxxxxxx >: > > > >> > > > >> Sure, also the scrubbing is happening on all the osds :S > > > >> > > > >> # ceph --cluster cephIB daemon osd.4 config diff > > > >> { > > > >> "diff": { > > > >> "current": { > > > >> "admin_socket": "\/var\/run\/ceph\/cephIB-osd.4.asok", > > > >> "auth_client_required": "cephx", > > > >> "filestore_fd_cache_size": "10240", > > > >> "filestore_journal_writeahead": "true", > > > >> "filestore_max_sync_interval": "10", > > > >> "filestore_merge_threshold": "40", > > > >> "filestore_op_threads": "20", > > > >> "filestore_queue_max_ops": "100000", > > > >> "filestore_split_multiple": "8", > > > >> "fsid": "a4bce51b-4d6b-4394-9737-3e4d9f5efed2", > > > >> "internal_safe_to_start_threads": "true", > > > >> "keyring": "\/var\/lib\/ceph\/osd\/cephIB-4\/keyring", > > > >> "leveldb_log": "", > > > >> "log_file": "\/var\/log\/ceph\/cephIB-osd.4.log", > > > >> "log_to_stderr": "false", > > > >> "mds_data": "\/var\/lib\/ceph\/mds\/cephIB-4", > > > >> "mon_cluster_log_file": > > > >> "default=\/var\/log\/ceph\/cephIB.$channel.log > > > >> cluster=\/var\/log\/ceph\/cephIB.log", > > > >> "mon_data": "\/var\/lib\/ceph\/mon\/cephIB-4", > > > >> "mon_debug_dump_location": > > > >> "\/var\/log\/ceph\/cephIB-osd.4.tdump", > > > >> "mon_host": "172.23.16.1,172.23.16.2,172.23.16.3", > > > >> "mon_initial_members": "cibm01, cibm02, cibm03", > > > >> "osd_data": "\/var\/lib\/ceph\/osd\/cephIB-4", > > > >> "osd_journal": "\/var\/lib\/ceph\/osd\/cephIB-4\/journal", > > > >> "osd_op_threads": "8", > > > >> "rgw_data": "\/var\/lib\/ceph\/radosgw\/cephIB-4", > > > >> "setgroup": "ceph", > > > >> "setuser": "ceph" > > > >> }, > > > >> "defaults": { > > > >> "admin_socket": "\/var\/run\/ceph\/ceph-osd.4.asok", > > > >> "auth_client_required": "cephx, none", > > > >> "filestore_fd_cache_size": "128", > > > >> "filestore_journal_writeahead": "false", > > > >> "filestore_max_sync_interval": "5", > > > >> "filestore_merge_threshold": "10", > > > >> "filestore_op_threads": "2", > > > >> "filestore_queue_max_ops": "50", > > > >> "filestore_split_multiple": "2", > > > >> "fsid": "00000000-0000-0000-0000-000000000000", > > > >> "internal_safe_to_start_threads": "false", > > > >> "keyring": > > > >> "\/etc\/ceph\/ceph.osd.4.keyring,\/etc\/ceph\/ceph.keyring,\/etc\/ceph\/keyring,\/etc\/ceph\/keyring.bin", > > > >> "leveldb_log": "\/dev\/null", > > > >> "log_file": "\/var\/log\/ceph\/ceph-osd.4.log", > > > >> "log_to_stderr": "true", > > > >> "mds_data": "\/var\/lib\/ceph\/mds\/ceph-4", > > > >> "mon_cluster_log_file": > > > >> "default=\/var\/log\/ceph\/ceph.$channel.log > > > >> cluster=\/var\/log\/ceph\/ceph.log", > > > >> "mon_data": "\/var\/lib\/ceph\/mon\/ceph-4", > > > >> "mon_debug_dump_location": > > > >> "\/var\/log\/ceph\/ceph-osd.4.tdump", > > > >> "mon_host": "", > > > >> "mon_initial_members": "", > > > >> "osd_data": "\/var\/lib\/ceph\/osd\/ceph-4", > > > >> "osd_journal": "\/var\/lib\/ceph\/osd\/ceph-4\/journal", > > > >> "osd_op_threads": "2", > > > >> "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-4", > > > >> "setgroup": "", > > > >> "setuser": "" > > > >> } > > > >> }, > > > >> "unknown": [] > > > >> } > > > >> > > > >> > > > >> Thanks a lot! > > > >> > > > >> Best, > > > >> > > > >> > > > >> German > > > >> > > > >> 2016-03-29 14:10 GMT-03:00 Samuel Just < sjust@xxxxxxxxxx >: > > > >>> > > > >>> That seems to be scrubbing pretty often. Can you attach a config diff > > > >>> from osd.4 (ceph daemon osd.4 config diff)? > > > >>> -Sam > > > >>> > > > >>> On Tue, Mar 29, 2016 at 9:30 AM, German Anders < ganders@xxxxxxxxxxxx > > > > >>> wrote: > > > >>> > Hi All, > > > >>> > > > > >>> > I've maybe a simple question, I've setup a new cluster with > > >>> > Infernalis > > > >>> > release, there's no IO going on at the cluster level and I'm > > >>> > receiving > > > >>> > a lot > > > >>> > of these messages: > > > >>> > > > > >>> > 2016-03-29 12:22:07.462818 mon.0 [INF] pgmap v158062: 8192 pgs: 8192 > > > >>> > active+clean; 20617 MB data, 46164 MB used, 52484 GB / 52529 GB avail > > > >>> > 2016-03-29 12:22:08.176684 osd.13 [INF] 0.d38 scrub starts > > > >>> > 2016-03-29 12:22:08.179841 osd.13 [INF] 0.d38 scrub ok > > > >>> > 2016-03-29 12:21:59.526355 osd.9 [INF] 0.8a6 scrub starts > > > >>> > 2016-03-29 12:21:59.529582 osd.9 [INF] 0.8a6 scrub ok > > > >>> > 2016-03-29 12:22:03.004107 osd.4 [INF] 0.38b scrub starts > > > >>> > 2016-03-29 12:22:03.007220 osd.4 [INF] 0.38b scrub ok > > > >>> > 2016-03-29 12:22:03.617706 osd.21 [INF] 0.525 scrub starts > > > >>> > 2016-03-29 12:22:03.621073 osd.21 [INF] 0.525 scrub ok > > > >>> > 2016-03-29 12:22:06.527264 osd.9 [INF] 0.8a6 scrub starts > > > >>> > 2016-03-29 12:22:06.529150 osd.9 [INF] 0.8a6 scrub ok > > > >>> > 2016-03-29 12:22:07.005628 osd.4 [INF] 0.38b scrub starts > > > >>> > 2016-03-29 12:22:07.009776 osd.4 [INF] 0.38b scrub ok > > > >>> > 2016-03-29 12:22:07.618191 osd.21 [INF] 0.525 scrub starts > > > >>> > 2016-03-29 12:22:07.621363 osd.21 [INF] 0.525 scrub ok > > > >>> > > > > >>> > > > > >>> > I mean, all the time, and AFAIK these is because the scrub operation > > >>> > is > > > >>> > like > > > >>> > an fsck on the object level, so this make me think that it's not a > > > >>> > normal > > > >>> > situation. Is there any command that I can run in order to check > > >>> > this? > > > >>> > > > > >>> > # ceph --cluster cephIB health detail > > > >>> > HEALTH_OK > > > >>> > > > > >>> > > > > >>> > Thanks in advance, > > > >>> > > > > >>> > Best, > > > >>> > > > > >>> > German > > > >>> > > > > >>> > _______________________________________________ > > > >>> > ceph-users mailing list > > > >>> > ceph-users@xxxxxxxxxxxxxx > > > >>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > >>> > > > > >> > > > >> > > > > > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@xxxxxxxxxxxxxx > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com