Re: Scrubbing a lot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Understood -- format 2 was promoted to the default image format starting with Infernalis (which not all users would have played with since it isn't LTS).  The defaults can be overridden via the command-line when creating new images or via the Ceph configuration file.

I'll let Ilya provide input on which kernels support image format 2, but from a quick peek on GitHub it looks like support was added around the v3.8 timeframe.

-- 

Jason Dillaman 

----- Original Message -----
> From: "Somnath Roy" <Somnath.Roy@xxxxxxxxxxx>
> To: "Jason Dillaman" <dillaman@xxxxxxxxxx>, "Stefan Lissmats" <stefan@xxxxxxxxxx>
> Cc: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
> Sent: Tuesday, March 29, 2016 8:24:17 PM
> Subject: RE:  Scrubbing a lot
> 
> We faced this issue too and figured out it in Jewel the default image
> creation was with format 2.
> Not sure if this is a good idea to change the default though as almost all
> the LTS releases are with older kernel and will face incompatibility issue.
> 
> Thanks & Regards
> Somnath
> 
> -----Original Message-----
> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of
> Jason Dillaman
> Sent: Tuesday, March 29, 2016 3:15 PM
> To: Stefan Lissmats
> Cc: ceph-users
> Subject: Re:  Scrubbing a lot
> 
> Image format 1 is still supported -- just trying to slowly move users off of
> it and onto image format 2 through lots of log message nagging.
> 
> --
> 
> Jason Dillaman
> 
> 
> ----- Original Message -----
> 
> > From: "Stefan Lissmats" <stefan@xxxxxxxxxx>
> > To: "German Anders" <ganders@xxxxxxxxxxxx>
> > Cc: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
> > Sent: Tuesday, March 29, 2016 4:55:02 PM
> > Subject: Re:  Scrubbing a lot
> 
> > Ok, i also got the warning but was able to use it anyway. Could be
> > blocked in the new release of Jewel. Probably the more correct answer
> > is to use the other answer (to use - - image-features layering) but
> > haven't tried that myself.
> 
> > Skickat från min Samsung-enhet
> 
> > -------- Originalmeddelande --------
> > Från: German Anders <ganders@xxxxxxxxxxxx>
> > Datum: 2016-03-29 22:48 (GMT+01:00)
> > Till: Stefan Lissmats <stefan@xxxxxxxxxx>
> > Kopia: Samuel Just <sjust@xxxxxxxxxx>, ceph-users
> > <ceph-users@xxxxxxxxxxxxxx>
> > Rubrik: Re:  Scrubbing a lot
> 
> > mmmm it seems that the image-format option is deprecated:
> 
> > # rbd --id cinder --cluster cephIB create e60host01v2 --size 100G
> > --image-format 1 --pool cinder-volumes -k
> > /etc/ceph/cephIB.client.cinder.keyring
> > rbd: image format 1 is deprecated
> 
> > # rbd --cluster cephIB info e60host01v2 --pool cinder-volumes
> > 2016-03-29 16:45:39.073198 7fb859eb7700 -1 librbd::image::OpenRequest:
> > RBD image format 1 is deprecated. Please copy this image to image format 2.
> > rbd image 'e60host01v2':
> > size 102400 MB in 25600 objects
> > order 22 (4096 kB objects)
> > block_name_prefix: rb.0.37d7.238e1f29
> > format: 1
> 
> > and the map operations still doesn't work :(
> 
> > # rbd --cluster cephIB map e60host01v2 --pool cinder-volumes -k
> > /etc/ceph/cephIB.client.cinder.keyring
> > rbd: sysfs write failed
> > rbd: map failed: (5) Input/output error
> 
> > also, I'm running kernel 3.19.0-39-generic
> 
> > German
> 
> > 2016-03-29 17:40 GMT-03:00 Stefan Lissmats < stefan@xxxxxxxxxx > :
> 
> > > I agrree. I ran in to the same issue and the error massage is not
> > > that clear.
> > > Mapping with the kernel rbd client (rbd map) needs a quite new
> > > kernel to handle the new image format. The work-around is to use - -
> > > image-format 1 when creating the image.
> >
> 
> > > -------- Originalmeddelande --------
> >
> > > Från: Samuel Just < sjust@xxxxxxxxxx >
> >
> > > Datum: 2016-03-29 22:24 (GMT+01:00)
> >
> > > Till: German Anders < ganders@xxxxxxxxxxxx >
> >
> > > Kopia: ceph-users < ceph-users@xxxxxxxxxxxxxx >
> >
> > > Rubrik: Re:  Scrubbing a lot
> >
> 
> > > Sounds like a version/compatibility thing. Are your rbd clients really
> > > old?
> >
> > > -Sam
> >
> 
> > > On Tue, Mar 29, 2016 at 1:19 PM, German Anders <
> > > ganders@xxxxxxxxxxxx >
> > > wrote:
> >
> > > > I've just upgrade to jewel, and the scrubbing seems to been
> > > > corrected...
> > > > but
> >
> > > > now I'm not able to map an rbd on a host (before I was able to),
> > > > basically
> >
> > > > I'm getting this error msg:
> >
> > > >
> >
> > > > rbd: sysfs write failed
> >
> > > > rbd: map failed: (5) Input/output error
> >
> > > >
> >
> > > > # rbd --cluster cephIB create host01 --size 102400 --pool
> > > > cinder-volumes -k
> >
> > > > /etc/ceph/cephIB.client.cinder.keyring
> >
> > > > # rbd --cluster cephIB map host01 --pool cinder-volumes -k
> >
> > > > /etc/ceph/cephIB.client.cinder.keyring
> >
> > > > rbd: sysfs write failed
> >
> > > > rbd: map failed: (5) Input/output error
> >
> > > >
> >
> > > > Any ideas? on the /etc/ceph directory on the host I've:
> >
> > > >
> >
> > > > -rw-r--r-- 1 ceph ceph 92 Nov 17 15:45 rbdmap
> >
> > > > -rw-r--r-- 1 ceph ceph 170 Dec 15 14:47 secret.xml
> >
> > > > -rw-r--r-- 1 ceph ceph 37 Dec 15 15:12 virsh-secret
> >
> > > > -rw-r--r-- 1 ceph ceph 0 Dec 15 15:12 virsh-secret-set
> >
> > > > -rw-r--r-- 1 ceph ceph 37 Dec 21 14:53 virsh-secretIB
> >
> > > > -rw-r--r-- 1 ceph ceph 0 Dec 21 14:53 virsh-secret-setIB
> >
> > > > -rw-r--r-- 1 ceph ceph 173 Dec 22 13:34 secretIB.xml
> >
> > > > -rw-r--r-- 1 ceph ceph 619 Dec 22 13:38 ceph.conf
> >
> > > > -rw-r--r-- 1 ceph ceph 72 Dec 23 09:51 ceph.client.cinder.keyring
> >
> > > > -rw-r--r-- 1 ceph ceph 63 Mar 28 09:03
> > > > cephIB.client.cinder.keyring
> >
> > > > -rw-r--r-- 1 ceph ceph 526 Mar 28 12:06 cephIB.conf
> >
> > > > -rw------- 1 ceph ceph 63 Mar 29 16:11 cephIB.client.admin.keyring
> >
> > > >
> >
> > > > Thanks in advance,
> >
> > > >
> >
> > > > Best,
> >
> > > >
> >
> > > > German
> >
> > > >
> >
> > > > 2016-03-29 14:45 GMT-03:00 German Anders < ganders@xxxxxxxxxxxx >:
> >
> > > >>
> >
> > > >> Sure, also the scrubbing is happening on all the osds :S
> >
> > > >>
> >
> > > >> # ceph --cluster cephIB daemon osd.4 config diff
> >
> > > >> {
> >
> > > >> "diff": {
> >
> > > >> "current": {
> >
> > > >> "admin_socket": "\/var\/run\/ceph\/cephIB-osd.4.asok",
> >
> > > >> "auth_client_required": "cephx",
> >
> > > >> "filestore_fd_cache_size": "10240",
> >
> > > >> "filestore_journal_writeahead": "true",
> >
> > > >> "filestore_max_sync_interval": "10",
> >
> > > >> "filestore_merge_threshold": "40",
> >
> > > >> "filestore_op_threads": "20",
> >
> > > >> "filestore_queue_max_ops": "100000",
> >
> > > >> "filestore_split_multiple": "8",
> >
> > > >> "fsid": "a4bce51b-4d6b-4394-9737-3e4d9f5efed2",
> >
> > > >> "internal_safe_to_start_threads": "true",
> >
> > > >> "keyring": "\/var\/lib\/ceph\/osd\/cephIB-4\/keyring",
> >
> > > >> "leveldb_log": "",
> >
> > > >> "log_file": "\/var\/log\/ceph\/cephIB-osd.4.log",
> >
> > > >> "log_to_stderr": "false",
> >
> > > >> "mds_data": "\/var\/lib\/ceph\/mds\/cephIB-4",
> >
> > > >> "mon_cluster_log_file":
> >
> > > >> "default=\/var\/log\/ceph\/cephIB.$channel.log
> >
> > > >> cluster=\/var\/log\/ceph\/cephIB.log",
> >
> > > >> "mon_data": "\/var\/lib\/ceph\/mon\/cephIB-4",
> >
> > > >> "mon_debug_dump_location":
> >
> > > >> "\/var\/log\/ceph\/cephIB-osd.4.tdump",
> >
> > > >> "mon_host": "172.23.16.1,172.23.16.2,172.23.16.3",
> >
> > > >> "mon_initial_members": "cibm01, cibm02, cibm03",
> >
> > > >> "osd_data": "\/var\/lib\/ceph\/osd\/cephIB-4",
> >
> > > >> "osd_journal": "\/var\/lib\/ceph\/osd\/cephIB-4\/journal",
> >
> > > >> "osd_op_threads": "8",
> >
> > > >> "rgw_data": "\/var\/lib\/ceph\/radosgw\/cephIB-4",
> >
> > > >> "setgroup": "ceph",
> >
> > > >> "setuser": "ceph"
> >
> > > >> },
> >
> > > >> "defaults": {
> >
> > > >> "admin_socket": "\/var\/run\/ceph\/ceph-osd.4.asok",
> >
> > > >> "auth_client_required": "cephx, none",
> >
> > > >> "filestore_fd_cache_size": "128",
> >
> > > >> "filestore_journal_writeahead": "false",
> >
> > > >> "filestore_max_sync_interval": "5",
> >
> > > >> "filestore_merge_threshold": "10",
> >
> > > >> "filestore_op_threads": "2",
> >
> > > >> "filestore_queue_max_ops": "50",
> >
> > > >> "filestore_split_multiple": "2",
> >
> > > >> "fsid": "00000000-0000-0000-0000-000000000000",
> >
> > > >> "internal_safe_to_start_threads": "false",
> >
> > > >> "keyring":
> >
> > > >> "\/etc\/ceph\/ceph.osd.4.keyring,\/etc\/ceph\/ceph.keyring,\/etc\
> > > >> /ceph\/keyring,\/etc\/ceph\/keyring.bin",
> >
> > > >> "leveldb_log": "\/dev\/null",
> >
> > > >> "log_file": "\/var\/log\/ceph\/ceph-osd.4.log",
> >
> > > >> "log_to_stderr": "true",
> >
> > > >> "mds_data": "\/var\/lib\/ceph\/mds\/ceph-4",
> >
> > > >> "mon_cluster_log_file":
> >
> > > >> "default=\/var\/log\/ceph\/ceph.$channel.log
> >
> > > >> cluster=\/var\/log\/ceph\/ceph.log",
> >
> > > >> "mon_data": "\/var\/lib\/ceph\/mon\/ceph-4",
> >
> > > >> "mon_debug_dump_location":
> >
> > > >> "\/var\/log\/ceph\/ceph-osd.4.tdump",
> >
> > > >> "mon_host": "",
> >
> > > >> "mon_initial_members": "",
> >
> > > >> "osd_data": "\/var\/lib\/ceph\/osd\/ceph-4",
> >
> > > >> "osd_journal": "\/var\/lib\/ceph\/osd\/ceph-4\/journal",
> >
> > > >> "osd_op_threads": "2",
> >
> > > >> "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-4",
> >
> > > >> "setgroup": "",
> >
> > > >> "setuser": ""
> >
> > > >> }
> >
> > > >> },
> >
> > > >> "unknown": []
> >
> > > >> }
> >
> > > >>
> >
> > > >>
> >
> > > >> Thanks a lot!
> >
> > > >>
> >
> > > >> Best,
> >
> > > >>
> >
> > > >>
> >
> > > >> German
> >
> > > >>
> >
> > > >> 2016-03-29 14:10 GMT-03:00 Samuel Just < sjust@xxxxxxxxxx >:
> >
> > > >>>
> >
> > > >>> That seems to be scrubbing pretty often. Can you attach a config
> > > >>> diff
> >
> > > >>> from osd.4 (ceph daemon osd.4 config diff)?
> >
> > > >>> -Sam
> >
> > > >>>
> >
> > > >>> On Tue, Mar 29, 2016 at 9:30 AM, German Anders <
> > > >>> ganders@xxxxxxxxxxxx >
> >
> > > >>> wrote:
> >
> > > >>> > Hi All,
> >
> > > >>> >
> >
> > > >>> > I've maybe a simple question, I've setup a new cluster with
> > > >>> > Infernalis
> >
> > > >>> > release, there's no IO going on at the cluster level and I'm
> > > >>> > receiving
> >
> > > >>> > a lot
> >
> > > >>> > of these messages:
> >
> > > >>> >
> >
> > > >>> > 2016-03-29 12:22:07.462818 mon.0 [INF] pgmap v158062: 8192
> > > >>> > pgs: 8192
> >
> > > >>> > active+clean; 20617 MB data, 46164 MB used, 52484 GB / 52529
> > > >>> > active+GB avail
> >
> > > >>> > 2016-03-29 12:22:08.176684 osd.13 [INF] 0.d38 scrub starts
> >
> > > >>> > 2016-03-29 12:22:08.179841 osd.13 [INF] 0.d38 scrub ok
> >
> > > >>> > 2016-03-29 12:21:59.526355 osd.9 [INF] 0.8a6 scrub starts
> >
> > > >>> > 2016-03-29 12:21:59.529582 osd.9 [INF] 0.8a6 scrub ok
> >
> > > >>> > 2016-03-29 12:22:03.004107 osd.4 [INF] 0.38b scrub starts
> >
> > > >>> > 2016-03-29 12:22:03.007220 osd.4 [INF] 0.38b scrub ok
> >
> > > >>> > 2016-03-29 12:22:03.617706 osd.21 [INF] 0.525 scrub starts
> >
> > > >>> > 2016-03-29 12:22:03.621073 osd.21 [INF] 0.525 scrub ok
> >
> > > >>> > 2016-03-29 12:22:06.527264 osd.9 [INF] 0.8a6 scrub starts
> >
> > > >>> > 2016-03-29 12:22:06.529150 osd.9 [INF] 0.8a6 scrub ok
> >
> > > >>> > 2016-03-29 12:22:07.005628 osd.4 [INF] 0.38b scrub starts
> >
> > > >>> > 2016-03-29 12:22:07.009776 osd.4 [INF] 0.38b scrub ok
> >
> > > >>> > 2016-03-29 12:22:07.618191 osd.21 [INF] 0.525 scrub starts
> >
> > > >>> > 2016-03-29 12:22:07.621363 osd.21 [INF] 0.525 scrub ok
> >
> > > >>> >
> >
> > > >>> >
> >
> > > >>> > I mean, all the time, and AFAIK these is because the scrub
> > > >>> > operation is
> >
> > > >>> > like
> >
> > > >>> > an fsck on the object level, so this make me think that it's
> > > >>> > not a
> >
> > > >>> > normal
> >
> > > >>> > situation. Is there any command that I can run in order to
> > > >>> > check this?
> >
> > > >>> >
> >
> > > >>> > # ceph --cluster cephIB health detail
> >
> > > >>> > HEALTH_OK
> >
> > > >>> >
> >
> > > >>> >
> >
> > > >>> > Thanks in advance,
> >
> > > >>> >
> >
> > > >>> > Best,
> >
> > > >>> >
> >
> > > >>> > German
> >
> > > >>> >
> >
> > > >>> > _______________________________________________
> >
> > > >>> > ceph-users mailing list
> >
> > > >>> > ceph-users@xxxxxxxxxxxxxx
> >
> > > >>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> > > >>> >
> >
> > > >>
> >
> > > >>
> >
> > > >
> >
> > > _______________________________________________
> >
> > > ceph-users mailing list
> >
> > > ceph-users@xxxxxxxxxxxxxx
> >
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> 
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> PLEASE NOTE: The information contained in this electronic mail message is
> intended only for the use of the designated recipient(s) named above. If the
> reader of this message is not the intended recipient, you are hereby
> notified that you have received this message in error and that any review,
> dissemination, distribution, or copying of this message is strictly
> prohibited. If you have received this communication in error, please notify
> the sender by telephone or e-mail (as shown above) immediately and destroy
> any and all copies of this message in your possession (whether hard copies
> or electronically stored copies).
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux