Sure, also the scrubbing is happening on all the osds :S
# ceph --cluster cephIB daemon osd.4 config diff
{
"diff": {
"current": {
"admin_socket": "\/var\/run\/ceph\/cephIB-osd.4.asok",
"auth_client_required": "cephx",
"filestore_fd_cache_size": "10240",
"filestore_journal_writeahead": "true",
"filestore_max_sync_interval": "10",
"filestore_merge_threshold": "40",
"filestore_op_threads": "20",
"filestore_queue_max_ops": "100000",
"filestore_split_multiple": "8",
"fsid": "a4bce51b-4d6b-4394-9737-3e4d9f5efed2",
"internal_safe_to_start_threads": "true",
"keyring": "\/var\/lib\/ceph\/osd\/cephIB-4\/keyring",
"leveldb_log": "",
"log_file": "\/var\/log\/ceph\/cephIB-osd.4.log",
"log_to_stderr": "false",
"mds_data": "\/var\/lib\/ceph\/mds\/cephIB-4",
"mon_cluster_log_file": "default=\/var\/log\/ceph\/cephIB.$channel.log cluster=\/var\/log\/ceph\/cephIB.log",
"mon_data": "\/var\/lib\/ceph\/mon\/cephIB-4",
"mon_debug_dump_location": "\/var\/log\/ceph\/cephIB-osd.4.tdump",
"mon_host": "172.23.16.1,172.23.16.2,172.23.16.3",
"mon_initial_members": "cibm01, cibm02, cibm03",
"osd_data": "\/var\/lib\/ceph\/osd\/cephIB-4",
"osd_journal": "\/var\/lib\/ceph\/osd\/cephIB-4\/journal",
"osd_op_threads": "8",
"rgw_data": "\/var\/lib\/ceph\/radosgw\/cephIB-4",
"setgroup": "ceph",
"setuser": "ceph"
},
"defaults": {
"admin_socket": "\/var\/run\/ceph\/ceph-osd.4.asok",
"auth_client_required": "cephx, none",
"filestore_fd_cache_size": "128",
"filestore_journal_writeahead": "false",
"filestore_max_sync_interval": "5",
"filestore_merge_threshold": "10",
"filestore_op_threads": "2",
"filestore_queue_max_ops": "50",
"filestore_split_multiple": "2",
"fsid": "00000000-0000-0000-0000-000000000000",
"internal_safe_to_start_threads": "false",
"keyring": "\/etc\/ceph\/ceph.osd.4.keyring,\/etc\/ceph\/ceph.keyring,\/etc\/ceph\/keyring,\/etc\/ceph\/keyring.bin",
"leveldb_log": "\/dev\/null",
"log_file": "\/var\/log\/ceph\/ceph-osd.4.log",
"log_to_stderr": "true",
"mds_data": "\/var\/lib\/ceph\/mds\/ceph-4",
"mon_cluster_log_file": "default=\/var\/log\/ceph\/ceph.$channel.log cluster=\/var\/log\/ceph\/ceph.log",
"mon_data": "\/var\/lib\/ceph\/mon\/ceph-4",
"mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.4.tdump",
"mon_host": "",
"mon_initial_members": "",
"osd_data": "\/var\/lib\/ceph\/osd\/ceph-4",
"osd_journal": "\/var\/lib\/ceph\/osd\/ceph-4\/journal",
"osd_op_threads": "2",
"rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-4",
"setgroup": "",
"setuser": ""
}
},
"unknown": []
}
{
"diff": {
"current": {
"admin_socket": "\/var\/run\/ceph\/cephIB-osd.4.asok",
"auth_client_required": "cephx",
"filestore_fd_cache_size": "10240",
"filestore_journal_writeahead": "true",
"filestore_max_sync_interval": "10",
"filestore_merge_threshold": "40",
"filestore_op_threads": "20",
"filestore_queue_max_ops": "100000",
"filestore_split_multiple": "8",
"fsid": "a4bce51b-4d6b-4394-9737-3e4d9f5efed2",
"internal_safe_to_start_threads": "true",
"keyring": "\/var\/lib\/ceph\/osd\/cephIB-4\/keyring",
"leveldb_log": "",
"log_file": "\/var\/log\/ceph\/cephIB-osd.4.log",
"log_to_stderr": "false",
"mds_data": "\/var\/lib\/ceph\/mds\/cephIB-4",
"mon_cluster_log_file": "default=\/var\/log\/ceph\/cephIB.$channel.log cluster=\/var\/log\/ceph\/cephIB.log",
"mon_data": "\/var\/lib\/ceph\/mon\/cephIB-4",
"mon_debug_dump_location": "\/var\/log\/ceph\/cephIB-osd.4.tdump",
"mon_host": "172.23.16.1,172.23.16.2,172.23.16.3",
"mon_initial_members": "cibm01, cibm02, cibm03",
"osd_data": "\/var\/lib\/ceph\/osd\/cephIB-4",
"osd_journal": "\/var\/lib\/ceph\/osd\/cephIB-4\/journal",
"osd_op_threads": "8",
"rgw_data": "\/var\/lib\/ceph\/radosgw\/cephIB-4",
"setgroup": "ceph",
"setuser": "ceph"
},
"defaults": {
"admin_socket": "\/var\/run\/ceph\/ceph-osd.4.asok",
"auth_client_required": "cephx, none",
"filestore_fd_cache_size": "128",
"filestore_journal_writeahead": "false",
"filestore_max_sync_interval": "5",
"filestore_merge_threshold": "10",
"filestore_op_threads": "2",
"filestore_queue_max_ops": "50",
"filestore_split_multiple": "2",
"fsid": "00000000-0000-0000-0000-000000000000",
"internal_safe_to_start_threads": "false",
"keyring": "\/etc\/ceph\/ceph.osd.4.keyring,\/etc\/ceph\/ceph.keyring,\/etc\/ceph\/keyring,\/etc\/ceph\/keyring.bin",
"leveldb_log": "\/dev\/null",
"log_file": "\/var\/log\/ceph\/ceph-osd.4.log",
"log_to_stderr": "true",
"mds_data": "\/var\/lib\/ceph\/mds\/ceph-4",
"mon_cluster_log_file": "default=\/var\/log\/ceph\/ceph.$channel.log cluster=\/var\/log\/ceph\/ceph.log",
"mon_data": "\/var\/lib\/ceph\/mon\/ceph-4",
"mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.4.tdump",
"mon_host": "",
"mon_initial_members": "",
"osd_data": "\/var\/lib\/ceph\/osd\/ceph-4",
"osd_journal": "\/var\/lib\/ceph\/osd\/ceph-4\/journal",
"osd_op_threads": "2",
"rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-4",
"setgroup": "",
"setuser": ""
}
},
"unknown": []
}
Thanks a lot!
Best,
German
2016-03-29 14:10 GMT-03:00 Samuel Just <sjust@xxxxxxxxxx>:
That seems to be scrubbing pretty often. Can you attach a config diff
from osd.4 (ceph daemon osd.4 config diff)?
-Sam
> _______________________________________________
On Tue, Mar 29, 2016 at 9:30 AM, German Anders <ganders@xxxxxxxxxxxx> wrote:
> Hi All,
>
> I've maybe a simple question, I've setup a new cluster with Infernalis
> release, there's no IO going on at the cluster level and I'm receiving a lot
> of these messages:
>
> 2016-03-29 12:22:07.462818 mon.0 [INF] pgmap v158062: 8192 pgs: 8192
> active+clean; 20617 MB data, 46164 MB used, 52484 GB / 52529 GB avail
> 2016-03-29 12:22:08.176684 osd.13 [INF] 0.d38 scrub starts
> 2016-03-29 12:22:08.179841 osd.13 [INF] 0.d38 scrub ok
> 2016-03-29 12:21:59.526355 osd.9 [INF] 0.8a6 scrub starts
> 2016-03-29 12:21:59.529582 osd.9 [INF] 0.8a6 scrub ok
> 2016-03-29 12:22:03.004107 osd.4 [INF] 0.38b scrub starts
> 2016-03-29 12:22:03.007220 osd.4 [INF] 0.38b scrub ok
> 2016-03-29 12:22:03.617706 osd.21 [INF] 0.525 scrub starts
> 2016-03-29 12:22:03.621073 osd.21 [INF] 0.525 scrub ok
> 2016-03-29 12:22:06.527264 osd.9 [INF] 0.8a6 scrub starts
> 2016-03-29 12:22:06.529150 osd.9 [INF] 0.8a6 scrub ok
> 2016-03-29 12:22:07.005628 osd.4 [INF] 0.38b scrub starts
> 2016-03-29 12:22:07.009776 osd.4 [INF] 0.38b scrub ok
> 2016-03-29 12:22:07.618191 osd.21 [INF] 0.525 scrub starts
> 2016-03-29 12:22:07.621363 osd.21 [INF] 0.525 scrub ok
>
>
> I mean, all the time, and AFAIK these is because the scrub operation is like
> an fsck on the object level, so this make me think that it's not a normal
> situation. Is there any command that I can run in order to check this?
>
> # ceph --cluster cephIB health detail
> HEALTH_OK
>
>
> Thanks in advance,
>
> Best,
>
> German
>
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com