Re: [cephadm] Found duplicate OSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Adam,

Getting the following error, not sure why it's not able to find it.

root@ceph1:~# ceph orch daemon redeploy mgr.ceph1.xmbvsb
Error EINVAL: Unable to find mgr.ceph1.xmbvsb daemon(s)

On Thu, Sep 1, 2022 at 5:57 PM Adam King <adking@xxxxxxxxxx> wrote:

> what happens if you run `ceph orch daemon redeploy mgr.ceph1.xmbvsb`?
>
> On Thu, Sep 1, 2022 at 5:12 PM Satish Patel <satish.txt@xxxxxxxxx> wrote:
>
>> Hi Adam,
>>
>> Here is requested output
>>
>> root@ceph1:~# ceph health detail
>> HEALTH_WARN 4 stray daemon(s) not managed by cephadm
>> [WRN] CEPHADM_STRAY_DAEMON: 4 stray daemon(s) not managed by cephadm
>>     stray daemon mon.ceph1 on host ceph1 not managed by cephadm
>>     stray daemon osd.0 on host ceph1 not managed by cephadm
>>     stray daemon osd.1 on host ceph1 not managed by cephadm
>>     stray daemon osd.4 on host ceph1 not managed by cephadm
>>
>>
>> root@ceph1:~# ceph orch host ls
>> HOST   ADDR         LABELS  STATUS
>> ceph1  10.73.0.192
>> ceph2  10.73.3.192  _admin
>> 2 hosts in cluster
>>
>>
>> My cephadm ls  saying mgr is in error state
>>
>> {
>>         "style": "cephadm:v1",
>>         "name": "mgr.ceph1.xmbvsb",
>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>         "systemd_unit":
>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb",
>>         "enabled": true,
>>         "state": "error",
>>         "container_id": null,
>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>         "container_image_id": null,
>>         "version": null,
>>         "started": null,
>>         "created": "2022-09-01T20:59:49.314347Z",
>>         "deployed": "2022-09-01T20:59:48.718347Z",
>>         "configured": "2022-09-01T20:59:49.314347Z"
>>     },
>>
>>
>> Getting error
>>
>> root@ceph1:~# cephadm unit --fsid f270ad9e-1f6f-11ed-b6f8-a539d87379ea
>> --name mgr.ceph1.xmbvsb start
>> stderr Job for
>> ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb.service
>> failed because the control process exited with error code.
>> stderr See "systemctl status
>> ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb.service" and
>> "journalctl -xe" for details.
>> Traceback (most recent call last):
>>   File "/usr/sbin/cephadm", line 6250, in <module>
>>     r = args.func()
>>   File "/usr/sbin/cephadm", line 1357, in _infer_fsid
>>     return func()
>>   File "/usr/sbin/cephadm", line 3727, in command_unit
>>     call_throws([
>>   File "/usr/sbin/cephadm", line 1119, in call_throws
>>     raise RuntimeError('Failed command: %s' % ' '.join(command))
>> RuntimeError: Failed command: systemctl start
>> ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb
>>
>>
>> How do I remove and re-deploy mgr?
>>
>> On Thu, Sep 1, 2022 at 4:54 PM Adam King <adking@xxxxxxxxxx> wrote:
>>
>>> cephadm deploys the containers with --rm so they will get removed if you
>>> stop them. As for getting the 2nd mgr back, if it still lists the 2nd one
>>> in `ceph orch ps` you should be able to do a `ceph orch daemon redeploy
>>> <mgr-daemon-name>` where <mgr-daemon-name> should match the name given in
>>> the orch ps output for the one that isn't actually up. If it isn't listed
>>> there, given you have a count of 2, cephadm should deploy another one. I do
>>> see in the orch ls output you posted that it says the mgr service has "2/2"
>>> running which implies it believes a 2nd mgr is present (and you would
>>> therefore be able to try the daemon redeploy if that daemon isn't actually
>>> there).
>>>
>>> Is it still reporting the duplicate osds in orch ps? I see in the
>>> cephadm ls output on ceph1 that osd.2 isn't being reported, which was
>>> reported as being on ceph1 in the orch ps output in your original message
>>> in this thread. I'm interested in what `ceph health detail` is reporting
>>> now as well, as it says there are 4 stray daemons. Also, the `ceph orch
>>> host ls` output just to get a better grasp of the topology of this cluster.
>>>
>>> On Thu, Sep 1, 2022 at 3:50 PM Satish Patel <satish.txt@xxxxxxxxx>
>>> wrote:
>>>
>>>> Adam,
>>>>
>>>> I have posted a question related to upgrading earlier and this thread
>>>> is related to that, I have opened a new one because I found that error in
>>>> logs and thought the upgrade may be stuck because of duplicate OSDs.
>>>>
>>>> root@ceph1:~# ls -l /var/lib/ceph/f270ad9e-1f6f-11ed-b6f8-a539d87379ea/
>>>> total 44
>>>> drwx------ 3 nobody nogroup 4096 Aug 19 05:37 alertmanager.ceph1
>>>> drwx------ 3    167     167 4096 Aug 19 05:36 crash
>>>> drwx------ 2    167     167 4096 Aug 19 05:37 crash.ceph1
>>>> drwx------ 4    998     996 4096 Aug 19 05:37 grafana.ceph1
>>>> drwx------ 2    167     167 4096 Aug 19 05:36 mgr.ceph1.xmbvsb
>>>> drwx------ 3    167     167 4096 Aug 19 05:36 mon.ceph1
>>>> drwx------ 2 nobody nogroup 4096 Aug 19 05:37 node-exporter.ceph1
>>>> drwx------ 2    167     167 4096 Aug 19 17:55 osd.0
>>>> drwx------ 2    167     167 4096 Aug 19 18:03 osd.1
>>>> drwx------ 2    167     167 4096 Aug 31 05:20 osd.4
>>>> drwx------ 4 nobody nogroup 4096 Aug 19 05:38 prometheus.ceph1
>>>>
>>>> Here is the output of cephadm ls
>>>>
>>>> root@ceph1:~# cephadm ls
>>>> [
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "alertmanager.ceph1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@alertmanager.ceph1",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "97403cf9799711461216b7f83e88c574da2b631c7c65233ebd82d8a216a48924",
>>>>         "container_image_name": "
>>>> quay.io/prometheus/alertmanager:v0.20.0",
>>>>         "container_image_id":
>>>> "0881eb8f169f5556a292b4e2c01d683172b12830a62a9225a98a8e206bb734f0",
>>>>         "version": "0.20.0",
>>>>         "started": "2022-08-19T16:59:02.461978Z",
>>>>         "created": "2022-08-19T03:37:16.403605Z",
>>>>         "deployed": "2022-08-19T03:37:15.815605Z",
>>>>         "configured": "2022-08-19T16:59:02.117607Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "grafana.ceph1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@grafana.ceph1",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "c7136aea8349a37dd9b320acd926c4bcbed95bc4549779e9580ed4290edc2117",
>>>>         "container_image_name": "quay.io/ceph/ceph-grafana:6.7.4",
>>>>         "container_image_id":
>>>> "557c83e11646f123a27b5e4b62ac6c45e7bb8b2e90d6044034d0db5b7019415c",
>>>>         "version": "6.7.4",
>>>>         "started": "2022-08-19T03:38:05.481992Z",
>>>>         "created": "2022-08-19T03:37:46.823604Z",
>>>>         "deployed": "2022-08-19T03:37:46.239604Z",
>>>>         "configured": "2022-08-19T03:38:05.163603Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "osd.1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit": "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@osd.1
>>>> ",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "51586b775bda0485c8b27b8401ac2430570e6f42cb7e12bae3eea05064f1fd20",
>>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>>         "container_image_id":
>>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>>         "version": "15.2.17",
>>>>         "started": "2022-08-19T16:03:10.612432Z",
>>>>         "created": "2022-08-19T16:03:09.765746Z",
>>>>         "deployed": "2022-08-19T16:03:09.141746Z",
>>>>         "configured": "2022-08-31T02:53:34.224643Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "prometheus.ceph1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@prometheus.ceph1",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "ba305236e5db9f2095b23b86a2340924909e9e8e54e5cdbe1d51c14dc4c8587a",
>>>>         "container_image_name": "quay.io/prometheus/prometheus:v2.18.1
>>>> ",
>>>>         "container_image_id":
>>>> "de242295e2257c37c8cadfd962369228f8f10b2d48a44259b65fef44ad4f6490",
>>>>         "version": "2.18.1",
>>>>         "started": "2022-08-19T16:59:03.538981Z",
>>>>         "created": "2022-08-19T03:38:01.567604Z",
>>>>         "deployed": "2022-08-19T03:38:00.983603Z",
>>>>         "configured": "2022-08-19T16:59:03.193607Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "node-exporter.ceph1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@node-exporter.ceph1",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "00bf3ad29cce79e905e8533648ef38cbd232990fa9616aff1c0020b7b66d0cc0",
>>>>         "container_image_name": "
>>>> quay.io/prometheus/node-exporter:v0.18.1",
>>>>         "container_image_id":
>>>> "e5a616e4b9cf68dfcad7782b78e118be4310022e874d52da85c55923fb615f87",
>>>>         "version": "0.18.1",
>>>>         "started": "2022-08-19T03:37:55.232032Z",
>>>>         "created": "2022-08-19T03:37:47.711604Z",
>>>>         "deployed": "2022-08-19T03:37:47.155604Z",
>>>>         "configured": "2022-08-19T03:37:47.711604Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "osd.0",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit": "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@osd.0
>>>> ",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "6b69046972dfbdb53665228258a15b13bc13a462ca4e066a4eca0cd593442d2d",
>>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>>         "container_image_id":
>>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>>         "version": "15.2.17",
>>>>         "started": "2022-08-19T15:55:20.580157Z",
>>>>         "created": "2022-08-19T15:55:19.725766Z",
>>>>         "deployed": "2022-08-19T15:55:19.125766Z",
>>>>         "configured": "2022-08-31T02:53:34.760643Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "crash.ceph1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@crash.ceph1",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "6bc56f478ccb96841fe86a540e284c175300b83dad9e906ae3230f22341c8293",
>>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>>         "container_image_id":
>>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>>         "version": "15.2.17",
>>>>         "started": "2022-08-19T03:37:17.660080Z",
>>>>         "created": "2022-08-19T03:37:17.559605Z",
>>>>         "deployed": "2022-08-19T03:37:16.991605Z",
>>>>         "configured": "2022-08-19T03:37:17.559605Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "mon.ceph1",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mon.ceph1",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "d0f03130491daebbe783c4990c6a4383d49e7a0e2bdf8c5d1eed012865e5d875",
>>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>>         "container_image_id":
>>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>>         "version": "15.2.17",
>>>>         "started": "2022-08-19T03:36:21.804129Z",
>>>>         "created": "2022-08-19T03:36:19.743608Z",
>>>>         "deployed": "2022-08-19T03:36:18.439608Z",
>>>>         "configured": "2022-08-19T03:38:05.931603Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "mgr.ceph1.xmbvsb",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit":
>>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb",
>>>>         "enabled": true,
>>>>         "state": "stopped",
>>>>         "container_id": null,
>>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>>         "container_image_id": null,
>>>>         "version": null,
>>>>         "started": null,
>>>>         "created": "2022-08-19T03:36:22.815608Z",
>>>>         "deployed": "2022-08-19T03:36:22.239608Z",
>>>>         "configured": "2022-08-19T03:38:06.487603Z"
>>>>     },
>>>>     {
>>>>         "style": "cephadm:v1",
>>>>         "name": "osd.4",
>>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>>         "systemd_unit": "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@osd.4
>>>> ",
>>>>         "enabled": true,
>>>>         "state": "running",
>>>>         "container_id":
>>>> "938840fe7fd0cb45cc26d077837c9847d7c7a7a68c7e1588d4bb4343c695a071",
>>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>>         "container_image_id":
>>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>>         "version": "15.2.17",
>>>>         "started": "2022-08-31T03:20:55.416219Z",
>>>>         "created": "2022-08-23T21:46:49.458533Z",
>>>>         "deployed": "2022-08-23T21:46:48.818533Z",
>>>>         "configured": "2022-08-31T02:53:41.196643Z"
>>>>     }
>>>> ]
>>>>
>>>>
>>>> I have noticed one more thing, I did docker stop <container_id_of_mgr>
>>>> on ceph1 node and now my mgr container disappeared, I can't see it anywhere
>>>> and not sure how do i bring back mgr because upgrade won't let me do
>>>> anything if i don't have two mgr instance.
>>>>
>>>> root@ceph1:~# ceph -s
>>>>   cluster:
>>>>     id:     f270ad9e-1f6f-11ed-b6f8-a539d87379ea
>>>>     health: HEALTH_WARN
>>>>             4 stray daemon(s) not managed by cephadm
>>>>
>>>>   services:
>>>>     mon: 1 daemons, quorum ceph1 (age 17h)
>>>>     mgr: ceph2.hmbdla(active, since 5h)
>>>>     osd: 6 osds: 6 up (since 40h), 6 in (since 8d)
>>>>
>>>>   data:
>>>>     pools:   6 pools, 161 pgs
>>>>     objects: 20.59k objects, 85 GiB
>>>>     usage:   174 GiB used, 826 GiB / 1000 GiB avail
>>>>     pgs:     161 active+clean
>>>>
>>>>   io:
>>>>     client:   0 B/s rd, 12 KiB/s wr, 0 op/s rd, 2 op/s wr
>>>>
>>>>   progress:
>>>>     Upgrade to quay.io/ceph/ceph:16.2.10 (0s)
>>>>       [............................]
>>>>
>>>> I can see mgr count:2 but not sure how do i bring it back
>>>>
>>>> root@ceph1:~# ceph orch ls
>>>> NAME                       PORTS        RUNNING  REFRESHED  AGE
>>>>  PLACEMENT
>>>> alertmanager               ?:9093,9094      1/1  20s ago    13d  count:1
>>>> crash                                       2/2  20s ago    13d  *
>>>> grafana                    ?:3000           1/1  20s ago    13d  count:1
>>>> mgr                                         2/2  20s ago    13d  count:2
>>>> mon                                         0/5  -          13d
>>>>  <unmanaged>
>>>> node-exporter              ?:9100           2/2  20s ago    13d  *
>>>> osd                                           6  20s ago    -
>>>>  <unmanaged>
>>>> osd.all-available-devices                     0  -          13d  *
>>>> osd.osd_spec_default                          0  -          8d   *
>>>> prometheus                 ?:9095           1/1  20s ago    13d  count:1
>>>>
>>>> On Thu, Sep 1, 2022 at 12:28 PM Adam King <adking@xxxxxxxxxx> wrote:
>>>>
>>>>> Are there any extra directories in /var/lib/ceph or
>>>>> /var/lib/ceph/<fsid> that appear to be for those OSDs on that host? When
>>>>> cephadm builds the info it uses for "ceph orch ps" it's actually scraping
>>>>> those directories. The output of "cephadm ls" on the host with the
>>>>> duplicates could also potentially have some insights.
>>>>>
>>>>> On Thu, Sep 1, 2022 at 12:15 PM Satish Patel <satish.txt@xxxxxxxxx>
>>>>> wrote:
>>>>>
>>>>>> Folks,
>>>>>>
>>>>>> I am playing with cephadm and life was good until I started upgrading
>>>>>> from
>>>>>> octopus to pacific. My upgrade process stuck after upgrading mgr and
>>>>>> in
>>>>>> logs now i can see following error
>>>>>>
>>>>>> root@ceph1:~# ceph log last cephadm
>>>>>> 2022-09-01T14:40:45.739804+0000 mgr.ceph2.hmbdla (mgr.265806) 8 :
>>>>>> cephadm [INF] Deploying daemon grafana.ceph1 on ceph1
>>>>>> 2022-09-01T14:40:56.115693+0000 mgr.ceph2.hmbdla (mgr.265806) 14 :
>>>>>> cephadm [INF] Deploying daemon prometheus.ceph1 on ceph1
>>>>>> 2022-09-01T14:41:11.856725+0000 mgr.ceph2.hmbdla (mgr.265806) 25 :
>>>>>> cephadm [INF] Reconfiguring alertmanager.ceph1 (dependencies
>>>>>> changed)...
>>>>>> 2022-09-01T14:41:11.861535+0000 mgr.ceph2.hmbdla (mgr.265806) 26 :
>>>>>> cephadm [INF] Reconfiguring daemon alertmanager.ceph1 on ceph1
>>>>>> 2022-09-01T14:41:12.927852+0000 mgr.ceph2.hmbdla (mgr.265806) 27 :
>>>>>> cephadm [INF] Reconfiguring grafana.ceph1 (dependencies changed)...
>>>>>> 2022-09-01T14:41:12.940615+0000 mgr.ceph2.hmbdla (mgr.265806) 28 :
>>>>>> cephadm [INF] Reconfiguring daemon grafana.ceph1 on ceph1
>>>>>> 2022-09-01T14:41:14.056113+0000 mgr.ceph2.hmbdla (mgr.265806) 33 :
>>>>>> cephadm [INF] Found duplicate OSDs: osd.2 in status running on ceph1,
>>>>>> osd.2 in status running on ceph2
>>>>>> 2022-09-01T14:41:14.056437+0000 mgr.ceph2.hmbdla (mgr.265806) 34 :
>>>>>> cephadm [INF] Found duplicate OSDs: osd.5 in status running on ceph1,
>>>>>> osd.5 in status running on ceph2
>>>>>> 2022-09-01T14:41:14.056630+0000 mgr.ceph2.hmbdla (mgr.265806) 35 :
>>>>>> cephadm [INF] Found duplicate OSDs: osd.3 in status running on ceph1,
>>>>>> osd.3 in status running on ceph2
>>>>>>
>>>>>>
>>>>>> Not sure from where duplicate names came and how that happened. In
>>>>>> following output i can't see any duplication
>>>>>>
>>>>>> root@ceph1:~# ceph osd tree
>>>>>> ID  CLASS  WEIGHT   TYPE NAME       STATUS  REWEIGHT  PRI-AFF
>>>>>> -1         0.97656  root default
>>>>>> -3         0.48828      host ceph1
>>>>>>  4    hdd  0.09769          osd.4       up   1.00000  1.00000
>>>>>>  0    ssd  0.19530          osd.0       up   1.00000  1.00000
>>>>>>  1    ssd  0.19530          osd.1       up   1.00000  1.00000
>>>>>> -5         0.48828      host ceph2
>>>>>>  5    hdd  0.09769          osd.5       up   1.00000  1.00000
>>>>>>  2    ssd  0.19530          osd.2       up   1.00000  1.00000
>>>>>>  3    ssd  0.19530          osd.3       up   1.00000  1.00000
>>>>>>
>>>>>>
>>>>>> But same time i can see duplicate OSD number in ceph1 and ceph2
>>>>>>
>>>>>>
>>>>>> root@ceph1:~# ceph orch ps
>>>>>> NAME                 HOST   PORTS        STATUS         REFRESHED  AGE
>>>>>>  MEM USE  MEM LIM  VERSION  IMAGE ID      CONTAINER ID
>>>>>> alertmanager.ceph1   ceph1  *:9093,9094  running (20s)     2s ago  20s
>>>>>>    17.1M        -           ba2b418f427c  856a4fe641f1
>>>>>> alertmanager.ceph1   ceph2  *:9093,9094  running (20s)     3s ago  20s
>>>>>>    17.1M        -           ba2b418f427c  856a4fe641f1
>>>>>> crash.ceph2          ceph1               running (12d)     2s ago  12d
>>>>>>    10.0M        -  15.2.17  93146564743f  0a009254afb0
>>>>>> crash.ceph2          ceph2               running (12d)     3s ago  12d
>>>>>>    10.0M        -  15.2.17  93146564743f  0a009254afb0
>>>>>> grafana.ceph1        ceph1  *:3000       running (18s)     2s ago  19s
>>>>>>    47.9M        -  8.3.5    dad864ee21e9  7d7a70b8ab7f
>>>>>> grafana.ceph1        ceph2  *:3000       running (18s)     3s ago  19s
>>>>>>    47.9M        -  8.3.5    dad864ee21e9  7d7a70b8ab7f
>>>>>> mgr.ceph2.hmbdla     ceph1               running (13h)     2s ago  12d
>>>>>>     506M        -  16.2.10  0d668911f040  6274723c35f7
>>>>>> mgr.ceph2.hmbdla     ceph2               running (13h)     3s ago  12d
>>>>>>     506M        -  16.2.10  0d668911f040  6274723c35f7
>>>>>> node-exporter.ceph2  ceph1               running (91m)     2s ago  12d
>>>>>>    60.7M        -  0.18.1   e5a616e4b9cf  d0ba04bb977c
>>>>>> node-exporter.ceph2  ceph2               running (91m)     3s ago  12d
>>>>>>    60.7M        -  0.18.1   e5a616e4b9cf  d0ba04bb977c
>>>>>> osd.2                ceph1               running (12h)     2s ago  12d
>>>>>>     867M    4096M  15.2.17  93146564743f  e286fb1c6302
>>>>>> osd.2                ceph2               running (12h)     3s ago  12d
>>>>>>     867M    4096M  15.2.17  93146564743f  e286fb1c6302
>>>>>> osd.3                ceph1               running (12h)     2s ago  12d
>>>>>>     978M    4096M  15.2.17  93146564743f  d3ae5d9f694f
>>>>>> osd.3                ceph2               running (12h)     3s ago  12d
>>>>>>     978M    4096M  15.2.17  93146564743f  d3ae5d9f694f
>>>>>> osd.5                ceph1               running (12h)     2s ago   8d
>>>>>>     225M    4096M  15.2.17  93146564743f  405068fb474e
>>>>>> osd.5                ceph2               running (12h)     3s ago   8d
>>>>>>     225M    4096M  15.2.17  93146564743f  405068fb474e
>>>>>> prometheus.ceph1     ceph1  *:9095       running (8s)      2s ago   8s
>>>>>>    30.4M        -           514e6a882f6e  9031dbe30cae
>>>>>> prometheus.ceph1     ceph2  *:9095       running (8s)      3s ago   8s
>>>>>>    30.4M        -           514e6a882f6e  9031dbe30cae
>>>>>>
>>>>>>
>>>>>> Is this a bug or did I do something wrong? any workaround to get out
>>>>>> from this condition?
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>>>>
>>>>>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux