Re: Scrubbing a lot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



What's the kernel version?
-Sam

On Tue, Mar 29, 2016 at 1:33 PM, German Anders <ganders@xxxxxxxxxxxx> wrote:
> On the host:
>
> # ceph --cluster cephIB --version
> ceph version 10.1.0 (96ae8bd25f31862dbd5302f304ebf8bf1166aba6)
>
> # rbd --version
> ceph version 10.1.0 (96ae8bd25f31862dbd5302f304ebf8bf1166aba6)
>
> If I run the command without root or sudo the command failed with a
> Permission denied
>
> German
>
> 2016-03-29 17:24 GMT-03:00 Samuel Just <sjust@xxxxxxxxxx>:
>>
>> Or you needed to run it as root?
>> -Sam
>>
>> On Tue, Mar 29, 2016 at 1:24 PM, Samuel Just <sjust@xxxxxxxxxx> wrote:
>> > Sounds like a version/compatibility thing.  Are your rbd clients really
>> > old?
>> > -Sam
>> >
>> > On Tue, Mar 29, 2016 at 1:19 PM, German Anders <ganders@xxxxxxxxxxxx>
>> > wrote:
>> >> I've just upgrade to jewel, and the scrubbing seems to been
>> >> corrected... but
>> >> now I'm not able to map an rbd on a host (before I was able to),
>> >> basically
>> >> I'm getting this error msg:
>> >>
>> >> rbd: sysfs write failed
>> >> rbd: map failed: (5) Input/output error
>> >>
>> >> # rbd --cluster cephIB create host01 --size 102400 --pool
>> >> cinder-volumes -k
>> >> /etc/ceph/cephIB.client.cinder.keyring
>> >> # rbd --cluster cephIB map host01 --pool cinder-volumes -k
>> >> /etc/ceph/cephIB.client.cinder.keyring
>> >> rbd: sysfs write failed
>> >> rbd: map failed: (5) Input/output error
>> >>
>> >> Any ideas? on the /etc/ceph directory on the host I've:
>> >>
>> >> -rw-r--r-- 1 ceph ceph  92 Nov 17 15:45 rbdmap
>> >> -rw-r--r-- 1 ceph ceph 170 Dec 15 14:47 secret.xml
>> >> -rw-r--r-- 1 ceph ceph  37 Dec 15 15:12 virsh-secret
>> >> -rw-r--r-- 1 ceph ceph   0 Dec 15 15:12 virsh-secret-set
>> >> -rw-r--r-- 1 ceph ceph  37 Dec 21 14:53 virsh-secretIB
>> >> -rw-r--r-- 1 ceph ceph   0 Dec 21 14:53 virsh-secret-setIB
>> >> -rw-r--r-- 1 ceph ceph 173 Dec 22 13:34 secretIB.xml
>> >> -rw-r--r-- 1 ceph ceph 619 Dec 22 13:38 ceph.conf
>> >> -rw-r--r-- 1 ceph ceph  72 Dec 23 09:51 ceph.client.cinder.keyring
>> >> -rw-r--r-- 1 ceph ceph  63 Mar 28 09:03 cephIB.client.cinder.keyring
>> >> -rw-r--r-- 1 ceph ceph 526 Mar 28 12:06 cephIB.conf
>> >> -rw------- 1 ceph ceph  63 Mar 29 16:11 cephIB.client.admin.keyring
>> >>
>> >> Thanks in advance,
>> >>
>> >> Best,
>> >>
>> >> German
>> >>
>> >> 2016-03-29 14:45 GMT-03:00 German Anders <ganders@xxxxxxxxxxxx>:
>> >>>
>> >>> Sure, also the scrubbing is happening on all the osds :S
>> >>>
>> >>> # ceph --cluster cephIB daemon osd.4 config diff
>> >>> {
>> >>>     "diff": {
>> >>>         "current": {
>> >>>             "admin_socket": "\/var\/run\/ceph\/cephIB-osd.4.asok",
>> >>>             "auth_client_required": "cephx",
>> >>>             "filestore_fd_cache_size": "10240",
>> >>>             "filestore_journal_writeahead": "true",
>> >>>             "filestore_max_sync_interval": "10",
>> >>>             "filestore_merge_threshold": "40",
>> >>>             "filestore_op_threads": "20",
>> >>>             "filestore_queue_max_ops": "100000",
>> >>>             "filestore_split_multiple": "8",
>> >>>             "fsid": "a4bce51b-4d6b-4394-9737-3e4d9f5efed2",
>> >>>             "internal_safe_to_start_threads": "true",
>> >>>             "keyring": "\/var\/lib\/ceph\/osd\/cephIB-4\/keyring",
>> >>>             "leveldb_log": "",
>> >>>             "log_file": "\/var\/log\/ceph\/cephIB-osd.4.log",
>> >>>             "log_to_stderr": "false",
>> >>>             "mds_data": "\/var\/lib\/ceph\/mds\/cephIB-4",
>> >>>             "mon_cluster_log_file":
>> >>> "default=\/var\/log\/ceph\/cephIB.$channel.log
>> >>> cluster=\/var\/log\/ceph\/cephIB.log",
>> >>>             "mon_data": "\/var\/lib\/ceph\/mon\/cephIB-4",
>> >>>             "mon_debug_dump_location":
>> >>> "\/var\/log\/ceph\/cephIB-osd.4.tdump",
>> >>>             "mon_host": "172.23.16.1,172.23.16.2,172.23.16.3",
>> >>>             "mon_initial_members": "cibm01, cibm02, cibm03",
>> >>>             "osd_data": "\/var\/lib\/ceph\/osd\/cephIB-4",
>> >>>             "osd_journal": "\/var\/lib\/ceph\/osd\/cephIB-4\/journal",
>> >>>             "osd_op_threads": "8",
>> >>>             "rgw_data": "\/var\/lib\/ceph\/radosgw\/cephIB-4",
>> >>>             "setgroup": "ceph",
>> >>>             "setuser": "ceph"
>> >>>         },
>> >>>         "defaults": {
>> >>>             "admin_socket": "\/var\/run\/ceph\/ceph-osd.4.asok",
>> >>>             "auth_client_required": "cephx, none",
>> >>>             "filestore_fd_cache_size": "128",
>> >>>             "filestore_journal_writeahead": "false",
>> >>>             "filestore_max_sync_interval": "5",
>> >>>             "filestore_merge_threshold": "10",
>> >>>             "filestore_op_threads": "2",
>> >>>             "filestore_queue_max_ops": "50",
>> >>>             "filestore_split_multiple": "2",
>> >>>             "fsid": "00000000-0000-0000-0000-000000000000",
>> >>>             "internal_safe_to_start_threads": "false",
>> >>>             "keyring":
>> >>>
>> >>> "\/etc\/ceph\/ceph.osd.4.keyring,\/etc\/ceph\/ceph.keyring,\/etc\/ceph\/keyring,\/etc\/ceph\/keyring.bin",
>> >>>             "leveldb_log": "\/dev\/null",
>> >>>             "log_file": "\/var\/log\/ceph\/ceph-osd.4.log",
>> >>>             "log_to_stderr": "true",
>> >>>             "mds_data": "\/var\/lib\/ceph\/mds\/ceph-4",
>> >>>             "mon_cluster_log_file":
>> >>> "default=\/var\/log\/ceph\/ceph.$channel.log
>> >>> cluster=\/var\/log\/ceph\/ceph.log",
>> >>>             "mon_data": "\/var\/lib\/ceph\/mon\/ceph-4",
>> >>>             "mon_debug_dump_location":
>> >>> "\/var\/log\/ceph\/ceph-osd.4.tdump",
>> >>>             "mon_host": "",
>> >>>             "mon_initial_members": "",
>> >>>             "osd_data": "\/var\/lib\/ceph\/osd\/ceph-4",
>> >>>             "osd_journal": "\/var\/lib\/ceph\/osd\/ceph-4\/journal",
>> >>>             "osd_op_threads": "2",
>> >>>             "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-4",
>> >>>             "setgroup": "",
>> >>>             "setuser": ""
>> >>>         }
>> >>>     },
>> >>>     "unknown": []
>> >>> }
>> >>>
>> >>>
>> >>> Thanks a lot!
>> >>>
>> >>> Best,
>> >>>
>> >>>
>> >>> German
>> >>>
>> >>> 2016-03-29 14:10 GMT-03:00 Samuel Just <sjust@xxxxxxxxxx>:
>> >>>>
>> >>>> That seems to be scrubbing pretty often.  Can you attach a config
>> >>>> diff
>> >>>> from osd.4 (ceph daemon osd.4 config diff)?
>> >>>> -Sam
>> >>>>
>> >>>> On Tue, Mar 29, 2016 at 9:30 AM, German Anders <ganders@xxxxxxxxxxxx>
>> >>>> wrote:
>> >>>> > Hi All,
>> >>>> >
>> >>>> > I've maybe a simple question, I've setup a new cluster with
>> >>>> > Infernalis
>> >>>> > release, there's no IO going on at the cluster level and I'm
>> >>>> > receiving
>> >>>> > a lot
>> >>>> > of these messages:
>> >>>> >
>> >>>> > 2016-03-29 12:22:07.462818 mon.0 [INF] pgmap v158062: 8192 pgs:
>> >>>> > 8192
>> >>>> > active+clean; 20617 MB data, 46164 MB used, 52484 GB / 52529 GB
>> >>>> > avail
>> >>>> > 2016-03-29 12:22:08.176684 osd.13 [INF] 0.d38 scrub starts
>> >>>> > 2016-03-29 12:22:08.179841 osd.13 [INF] 0.d38 scrub ok
>> >>>> > 2016-03-29 12:21:59.526355 osd.9 [INF] 0.8a6 scrub starts
>> >>>> > 2016-03-29 12:21:59.529582 osd.9 [INF] 0.8a6 scrub ok
>> >>>> > 2016-03-29 12:22:03.004107 osd.4 [INF] 0.38b scrub starts
>> >>>> > 2016-03-29 12:22:03.007220 osd.4 [INF] 0.38b scrub ok
>> >>>> > 2016-03-29 12:22:03.617706 osd.21 [INF] 0.525 scrub starts
>> >>>> > 2016-03-29 12:22:03.621073 osd.21 [INF] 0.525 scrub ok
>> >>>> > 2016-03-29 12:22:06.527264 osd.9 [INF] 0.8a6 scrub starts
>> >>>> > 2016-03-29 12:22:06.529150 osd.9 [INF] 0.8a6 scrub ok
>> >>>> > 2016-03-29 12:22:07.005628 osd.4 [INF] 0.38b scrub starts
>> >>>> > 2016-03-29 12:22:07.009776 osd.4 [INF] 0.38b scrub ok
>> >>>> > 2016-03-29 12:22:07.618191 osd.21 [INF] 0.525 scrub starts
>> >>>> > 2016-03-29 12:22:07.621363 osd.21 [INF] 0.525 scrub ok
>> >>>> >
>> >>>> >
>> >>>> > I mean, all the time, and AFAIK these is because the scrub
>> >>>> > operation is
>> >>>> > like
>> >>>> > an fsck on the object level, so this make me think that it's not a
>> >>>> > normal
>> >>>> > situation. Is there any command that I can run in order to check
>> >>>> > this?
>> >>>> >
>> >>>> > # ceph --cluster cephIB health detail
>> >>>> > HEALTH_OK
>> >>>> >
>> >>>> >
>> >>>> > Thanks in advance,
>> >>>> >
>> >>>> > Best,
>> >>>> >
>> >>>> > German
>> >>>> >
>> >>>> > _______________________________________________
>> >>>> > ceph-users mailing list
>> >>>> > ceph-users@xxxxxxxxxxxxxx
>> >>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >>>> >
>> >>>
>> >>>
>> >>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux