Re: [cephadm] Found duplicate OSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



what happens if you run `ceph orch daemon redeploy mgr.ceph1.xmbvsb`?

On Thu, Sep 1, 2022 at 5:12 PM Satish Patel <satish.txt@xxxxxxxxx> wrote:

> Hi Adam,
>
> Here is requested output
>
> root@ceph1:~# ceph health detail
> HEALTH_WARN 4 stray daemon(s) not managed by cephadm
> [WRN] CEPHADM_STRAY_DAEMON: 4 stray daemon(s) not managed by cephadm
>     stray daemon mon.ceph1 on host ceph1 not managed by cephadm
>     stray daemon osd.0 on host ceph1 not managed by cephadm
>     stray daemon osd.1 on host ceph1 not managed by cephadm
>     stray daemon osd.4 on host ceph1 not managed by cephadm
>
>
> root@ceph1:~# ceph orch host ls
> HOST   ADDR         LABELS  STATUS
> ceph1  10.73.0.192
> ceph2  10.73.3.192  _admin
> 2 hosts in cluster
>
>
> My cephadm ls  saying mgr is in error state
>
> {
>         "style": "cephadm:v1",
>         "name": "mgr.ceph1.xmbvsb",
>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>         "systemd_unit":
> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb",
>         "enabled": true,
>         "state": "error",
>         "container_id": null,
>         "container_image_name": "quay.io/ceph/ceph:v15",
>         "container_image_id": null,
>         "version": null,
>         "started": null,
>         "created": "2022-09-01T20:59:49.314347Z",
>         "deployed": "2022-09-01T20:59:48.718347Z",
>         "configured": "2022-09-01T20:59:49.314347Z"
>     },
>
>
> Getting error
>
> root@ceph1:~# cephadm unit --fsid f270ad9e-1f6f-11ed-b6f8-a539d87379ea
> --name mgr.ceph1.xmbvsb start
> stderr Job for
> ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb.service failed
> because the control process exited with error code.
> stderr See "systemctl status
> ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb.service" and
> "journalctl -xe" for details.
> Traceback (most recent call last):
>   File "/usr/sbin/cephadm", line 6250, in <module>
>     r = args.func()
>   File "/usr/sbin/cephadm", line 1357, in _infer_fsid
>     return func()
>   File "/usr/sbin/cephadm", line 3727, in command_unit
>     call_throws([
>   File "/usr/sbin/cephadm", line 1119, in call_throws
>     raise RuntimeError('Failed command: %s' % ' '.join(command))
> RuntimeError: Failed command: systemctl start
> ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb
>
>
> How do I remove and re-deploy mgr?
>
> On Thu, Sep 1, 2022 at 4:54 PM Adam King <adking@xxxxxxxxxx> wrote:
>
>> cephadm deploys the containers with --rm so they will get removed if you
>> stop them. As for getting the 2nd mgr back, if it still lists the 2nd one
>> in `ceph orch ps` you should be able to do a `ceph orch daemon redeploy
>> <mgr-daemon-name>` where <mgr-daemon-name> should match the name given in
>> the orch ps output for the one that isn't actually up. If it isn't listed
>> there, given you have a count of 2, cephadm should deploy another one. I do
>> see in the orch ls output you posted that it says the mgr service has "2/2"
>> running which implies it believes a 2nd mgr is present (and you would
>> therefore be able to try the daemon redeploy if that daemon isn't actually
>> there).
>>
>> Is it still reporting the duplicate osds in orch ps? I see in the cephadm
>> ls output on ceph1 that osd.2 isn't being reported, which was reported as
>> being on ceph1 in the orch ps output in your original message in this
>> thread. I'm interested in what `ceph health detail` is reporting now as
>> well, as it says there are 4 stray daemons. Also, the `ceph orch host ls`
>> output just to get a better grasp of the topology of this cluster.
>>
>> On Thu, Sep 1, 2022 at 3:50 PM Satish Patel <satish.txt@xxxxxxxxx> wrote:
>>
>>> Adam,
>>>
>>> I have posted a question related to upgrading earlier and this thread is
>>> related to that, I have opened a new one because I found that error in logs
>>> and thought the upgrade may be stuck because of duplicate OSDs.
>>>
>>> root@ceph1:~# ls -l /var/lib/ceph/f270ad9e-1f6f-11ed-b6f8-a539d87379ea/
>>> total 44
>>> drwx------ 3 nobody nogroup 4096 Aug 19 05:37 alertmanager.ceph1
>>> drwx------ 3    167     167 4096 Aug 19 05:36 crash
>>> drwx------ 2    167     167 4096 Aug 19 05:37 crash.ceph1
>>> drwx------ 4    998     996 4096 Aug 19 05:37 grafana.ceph1
>>> drwx------ 2    167     167 4096 Aug 19 05:36 mgr.ceph1.xmbvsb
>>> drwx------ 3    167     167 4096 Aug 19 05:36 mon.ceph1
>>> drwx------ 2 nobody nogroup 4096 Aug 19 05:37 node-exporter.ceph1
>>> drwx------ 2    167     167 4096 Aug 19 17:55 osd.0
>>> drwx------ 2    167     167 4096 Aug 19 18:03 osd.1
>>> drwx------ 2    167     167 4096 Aug 31 05:20 osd.4
>>> drwx------ 4 nobody nogroup 4096 Aug 19 05:38 prometheus.ceph1
>>>
>>> Here is the output of cephadm ls
>>>
>>> root@ceph1:~# cephadm ls
>>> [
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "alertmanager.ceph1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@alertmanager.ceph1",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "97403cf9799711461216b7f83e88c574da2b631c7c65233ebd82d8a216a48924",
>>>         "container_image_name": "quay.io/prometheus/alertmanager:v0.20.0
>>> ",
>>>         "container_image_id":
>>> "0881eb8f169f5556a292b4e2c01d683172b12830a62a9225a98a8e206bb734f0",
>>>         "version": "0.20.0",
>>>         "started": "2022-08-19T16:59:02.461978Z",
>>>         "created": "2022-08-19T03:37:16.403605Z",
>>>         "deployed": "2022-08-19T03:37:15.815605Z",
>>>         "configured": "2022-08-19T16:59:02.117607Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "grafana.ceph1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@grafana.ceph1",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "c7136aea8349a37dd9b320acd926c4bcbed95bc4549779e9580ed4290edc2117",
>>>         "container_image_name": "quay.io/ceph/ceph-grafana:6.7.4",
>>>         "container_image_id":
>>> "557c83e11646f123a27b5e4b62ac6c45e7bb8b2e90d6044034d0db5b7019415c",
>>>         "version": "6.7.4",
>>>         "started": "2022-08-19T03:38:05.481992Z",
>>>         "created": "2022-08-19T03:37:46.823604Z",
>>>         "deployed": "2022-08-19T03:37:46.239604Z",
>>>         "configured": "2022-08-19T03:38:05.163603Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "osd.1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit": "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@osd.1
>>> ",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "51586b775bda0485c8b27b8401ac2430570e6f42cb7e12bae3eea05064f1fd20",
>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>         "container_image_id":
>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>         "version": "15.2.17",
>>>         "started": "2022-08-19T16:03:10.612432Z",
>>>         "created": "2022-08-19T16:03:09.765746Z",
>>>         "deployed": "2022-08-19T16:03:09.141746Z",
>>>         "configured": "2022-08-31T02:53:34.224643Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "prometheus.ceph1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@prometheus.ceph1",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "ba305236e5db9f2095b23b86a2340924909e9e8e54e5cdbe1d51c14dc4c8587a",
>>>         "container_image_name": "quay.io/prometheus/prometheus:v2.18.1",
>>>         "container_image_id":
>>> "de242295e2257c37c8cadfd962369228f8f10b2d48a44259b65fef44ad4f6490",
>>>         "version": "2.18.1",
>>>         "started": "2022-08-19T16:59:03.538981Z",
>>>         "created": "2022-08-19T03:38:01.567604Z",
>>>         "deployed": "2022-08-19T03:38:00.983603Z",
>>>         "configured": "2022-08-19T16:59:03.193607Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "node-exporter.ceph1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@node-exporter.ceph1",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "00bf3ad29cce79e905e8533648ef38cbd232990fa9616aff1c0020b7b66d0cc0",
>>>         "container_image_name": "
>>> quay.io/prometheus/node-exporter:v0.18.1",
>>>         "container_image_id":
>>> "e5a616e4b9cf68dfcad7782b78e118be4310022e874d52da85c55923fb615f87",
>>>         "version": "0.18.1",
>>>         "started": "2022-08-19T03:37:55.232032Z",
>>>         "created": "2022-08-19T03:37:47.711604Z",
>>>         "deployed": "2022-08-19T03:37:47.155604Z",
>>>         "configured": "2022-08-19T03:37:47.711604Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "osd.0",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit": "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@osd.0
>>> ",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "6b69046972dfbdb53665228258a15b13bc13a462ca4e066a4eca0cd593442d2d",
>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>         "container_image_id":
>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>         "version": "15.2.17",
>>>         "started": "2022-08-19T15:55:20.580157Z",
>>>         "created": "2022-08-19T15:55:19.725766Z",
>>>         "deployed": "2022-08-19T15:55:19.125766Z",
>>>         "configured": "2022-08-31T02:53:34.760643Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "crash.ceph1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@crash.ceph1",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "6bc56f478ccb96841fe86a540e284c175300b83dad9e906ae3230f22341c8293",
>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>         "container_image_id":
>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>         "version": "15.2.17",
>>>         "started": "2022-08-19T03:37:17.660080Z",
>>>         "created": "2022-08-19T03:37:17.559605Z",
>>>         "deployed": "2022-08-19T03:37:16.991605Z",
>>>         "configured": "2022-08-19T03:37:17.559605Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "mon.ceph1",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mon.ceph1",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "d0f03130491daebbe783c4990c6a4383d49e7a0e2bdf8c5d1eed012865e5d875",
>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>         "container_image_id":
>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>         "version": "15.2.17",
>>>         "started": "2022-08-19T03:36:21.804129Z",
>>>         "created": "2022-08-19T03:36:19.743608Z",
>>>         "deployed": "2022-08-19T03:36:18.439608Z",
>>>         "configured": "2022-08-19T03:38:05.931603Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "mgr.ceph1.xmbvsb",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit":
>>> "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@mgr.ceph1.xmbvsb",
>>>         "enabled": true,
>>>         "state": "stopped",
>>>         "container_id": null,
>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>         "container_image_id": null,
>>>         "version": null,
>>>         "started": null,
>>>         "created": "2022-08-19T03:36:22.815608Z",
>>>         "deployed": "2022-08-19T03:36:22.239608Z",
>>>         "configured": "2022-08-19T03:38:06.487603Z"
>>>     },
>>>     {
>>>         "style": "cephadm:v1",
>>>         "name": "osd.4",
>>>         "fsid": "f270ad9e-1f6f-11ed-b6f8-a539d87379ea",
>>>         "systemd_unit": "ceph-f270ad9e-1f6f-11ed-b6f8-a539d87379ea@osd.4
>>> ",
>>>         "enabled": true,
>>>         "state": "running",
>>>         "container_id":
>>> "938840fe7fd0cb45cc26d077837c9847d7c7a7a68c7e1588d4bb4343c695a071",
>>>         "container_image_name": "quay.io/ceph/ceph:v15",
>>>         "container_image_id":
>>> "93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
>>>         "version": "15.2.17",
>>>         "started": "2022-08-31T03:20:55.416219Z",
>>>         "created": "2022-08-23T21:46:49.458533Z",
>>>         "deployed": "2022-08-23T21:46:48.818533Z",
>>>         "configured": "2022-08-31T02:53:41.196643Z"
>>>     }
>>> ]
>>>
>>>
>>> I have noticed one more thing, I did docker stop <container_id_of_mgr>
>>> on ceph1 node and now my mgr container disappeared, I can't see it anywhere
>>> and not sure how do i bring back mgr because upgrade won't let me do
>>> anything if i don't have two mgr instance.
>>>
>>> root@ceph1:~# ceph -s
>>>   cluster:
>>>     id:     f270ad9e-1f6f-11ed-b6f8-a539d87379ea
>>>     health: HEALTH_WARN
>>>             4 stray daemon(s) not managed by cephadm
>>>
>>>   services:
>>>     mon: 1 daemons, quorum ceph1 (age 17h)
>>>     mgr: ceph2.hmbdla(active, since 5h)
>>>     osd: 6 osds: 6 up (since 40h), 6 in (since 8d)
>>>
>>>   data:
>>>     pools:   6 pools, 161 pgs
>>>     objects: 20.59k objects, 85 GiB
>>>     usage:   174 GiB used, 826 GiB / 1000 GiB avail
>>>     pgs:     161 active+clean
>>>
>>>   io:
>>>     client:   0 B/s rd, 12 KiB/s wr, 0 op/s rd, 2 op/s wr
>>>
>>>   progress:
>>>     Upgrade to quay.io/ceph/ceph:16.2.10 (0s)
>>>       [............................]
>>>
>>> I can see mgr count:2 but not sure how do i bring it back
>>>
>>> root@ceph1:~# ceph orch ls
>>> NAME                       PORTS        RUNNING  REFRESHED  AGE
>>>  PLACEMENT
>>> alertmanager               ?:9093,9094      1/1  20s ago    13d  count:1
>>> crash                                       2/2  20s ago    13d  *
>>> grafana                    ?:3000           1/1  20s ago    13d  count:1
>>> mgr                                         2/2  20s ago    13d  count:2
>>> mon                                         0/5  -          13d
>>>  <unmanaged>
>>> node-exporter              ?:9100           2/2  20s ago    13d  *
>>> osd                                           6  20s ago    -
>>>  <unmanaged>
>>> osd.all-available-devices                     0  -          13d  *
>>> osd.osd_spec_default                          0  -          8d   *
>>> prometheus                 ?:9095           1/1  20s ago    13d  count:1
>>>
>>> On Thu, Sep 1, 2022 at 12:28 PM Adam King <adking@xxxxxxxxxx> wrote:
>>>
>>>> Are there any extra directories in /var/lib/ceph or
>>>> /var/lib/ceph/<fsid> that appear to be for those OSDs on that host? When
>>>> cephadm builds the info it uses for "ceph orch ps" it's actually scraping
>>>> those directories. The output of "cephadm ls" on the host with the
>>>> duplicates could also potentially have some insights.
>>>>
>>>> On Thu, Sep 1, 2022 at 12:15 PM Satish Patel <satish.txt@xxxxxxxxx>
>>>> wrote:
>>>>
>>>>> Folks,
>>>>>
>>>>> I am playing with cephadm and life was good until I started upgrading
>>>>> from
>>>>> octopus to pacific. My upgrade process stuck after upgrading mgr and in
>>>>> logs now i can see following error
>>>>>
>>>>> root@ceph1:~# ceph log last cephadm
>>>>> 2022-09-01T14:40:45.739804+0000 mgr.ceph2.hmbdla (mgr.265806) 8 :
>>>>> cephadm [INF] Deploying daemon grafana.ceph1 on ceph1
>>>>> 2022-09-01T14:40:56.115693+0000 mgr.ceph2.hmbdla (mgr.265806) 14 :
>>>>> cephadm [INF] Deploying daemon prometheus.ceph1 on ceph1
>>>>> 2022-09-01T14:41:11.856725+0000 mgr.ceph2.hmbdla (mgr.265806) 25 :
>>>>> cephadm [INF] Reconfiguring alertmanager.ceph1 (dependencies
>>>>> changed)...
>>>>> 2022-09-01T14:41:11.861535+0000 mgr.ceph2.hmbdla (mgr.265806) 26 :
>>>>> cephadm [INF] Reconfiguring daemon alertmanager.ceph1 on ceph1
>>>>> 2022-09-01T14:41:12.927852+0000 mgr.ceph2.hmbdla (mgr.265806) 27 :
>>>>> cephadm [INF] Reconfiguring grafana.ceph1 (dependencies changed)...
>>>>> 2022-09-01T14:41:12.940615+0000 mgr.ceph2.hmbdla (mgr.265806) 28 :
>>>>> cephadm [INF] Reconfiguring daemon grafana.ceph1 on ceph1
>>>>> 2022-09-01T14:41:14.056113+0000 mgr.ceph2.hmbdla (mgr.265806) 33 :
>>>>> cephadm [INF] Found duplicate OSDs: osd.2 in status running on ceph1,
>>>>> osd.2 in status running on ceph2
>>>>> 2022-09-01T14:41:14.056437+0000 mgr.ceph2.hmbdla (mgr.265806) 34 :
>>>>> cephadm [INF] Found duplicate OSDs: osd.5 in status running on ceph1,
>>>>> osd.5 in status running on ceph2
>>>>> 2022-09-01T14:41:14.056630+0000 mgr.ceph2.hmbdla (mgr.265806) 35 :
>>>>> cephadm [INF] Found duplicate OSDs: osd.3 in status running on ceph1,
>>>>> osd.3 in status running on ceph2
>>>>>
>>>>>
>>>>> Not sure from where duplicate names came and how that happened. In
>>>>> following output i can't see any duplication
>>>>>
>>>>> root@ceph1:~# ceph osd tree
>>>>> ID  CLASS  WEIGHT   TYPE NAME       STATUS  REWEIGHT  PRI-AFF
>>>>> -1         0.97656  root default
>>>>> -3         0.48828      host ceph1
>>>>>  4    hdd  0.09769          osd.4       up   1.00000  1.00000
>>>>>  0    ssd  0.19530          osd.0       up   1.00000  1.00000
>>>>>  1    ssd  0.19530          osd.1       up   1.00000  1.00000
>>>>> -5         0.48828      host ceph2
>>>>>  5    hdd  0.09769          osd.5       up   1.00000  1.00000
>>>>>  2    ssd  0.19530          osd.2       up   1.00000  1.00000
>>>>>  3    ssd  0.19530          osd.3       up   1.00000  1.00000
>>>>>
>>>>>
>>>>> But same time i can see duplicate OSD number in ceph1 and ceph2
>>>>>
>>>>>
>>>>> root@ceph1:~# ceph orch ps
>>>>> NAME                 HOST   PORTS        STATUS         REFRESHED  AGE
>>>>>  MEM USE  MEM LIM  VERSION  IMAGE ID      CONTAINER ID
>>>>> alertmanager.ceph1   ceph1  *:9093,9094  running (20s)     2s ago  20s
>>>>>    17.1M        -           ba2b418f427c  856a4fe641f1
>>>>> alertmanager.ceph1   ceph2  *:9093,9094  running (20s)     3s ago  20s
>>>>>    17.1M        -           ba2b418f427c  856a4fe641f1
>>>>> crash.ceph2          ceph1               running (12d)     2s ago  12d
>>>>>    10.0M        -  15.2.17  93146564743f  0a009254afb0
>>>>> crash.ceph2          ceph2               running (12d)     3s ago  12d
>>>>>    10.0M        -  15.2.17  93146564743f  0a009254afb0
>>>>> grafana.ceph1        ceph1  *:3000       running (18s)     2s ago  19s
>>>>>    47.9M        -  8.3.5    dad864ee21e9  7d7a70b8ab7f
>>>>> grafana.ceph1        ceph2  *:3000       running (18s)     3s ago  19s
>>>>>    47.9M        -  8.3.5    dad864ee21e9  7d7a70b8ab7f
>>>>> mgr.ceph2.hmbdla     ceph1               running (13h)     2s ago  12d
>>>>>     506M        -  16.2.10  0d668911f040  6274723c35f7
>>>>> mgr.ceph2.hmbdla     ceph2               running (13h)     3s ago  12d
>>>>>     506M        -  16.2.10  0d668911f040  6274723c35f7
>>>>> node-exporter.ceph2  ceph1               running (91m)     2s ago  12d
>>>>>    60.7M        -  0.18.1   e5a616e4b9cf  d0ba04bb977c
>>>>> node-exporter.ceph2  ceph2               running (91m)     3s ago  12d
>>>>>    60.7M        -  0.18.1   e5a616e4b9cf  d0ba04bb977c
>>>>> osd.2                ceph1               running (12h)     2s ago  12d
>>>>>     867M    4096M  15.2.17  93146564743f  e286fb1c6302
>>>>> osd.2                ceph2               running (12h)     3s ago  12d
>>>>>     867M    4096M  15.2.17  93146564743f  e286fb1c6302
>>>>> osd.3                ceph1               running (12h)     2s ago  12d
>>>>>     978M    4096M  15.2.17  93146564743f  d3ae5d9f694f
>>>>> osd.3                ceph2               running (12h)     3s ago  12d
>>>>>     978M    4096M  15.2.17  93146564743f  d3ae5d9f694f
>>>>> osd.5                ceph1               running (12h)     2s ago   8d
>>>>>     225M    4096M  15.2.17  93146564743f  405068fb474e
>>>>> osd.5                ceph2               running (12h)     3s ago   8d
>>>>>     225M    4096M  15.2.17  93146564743f  405068fb474e
>>>>> prometheus.ceph1     ceph1  *:9095       running (8s)      2s ago   8s
>>>>>    30.4M        -           514e6a882f6e  9031dbe30cae
>>>>> prometheus.ceph1     ceph2  *:9095       running (8s)      3s ago   8s
>>>>>    30.4M        -           514e6a882f6e  9031dbe30cae
>>>>>
>>>>>
>>>>> Is this a bug or did I do something wrong? any workaround to get out
>>>>> from this condition?
>>>>> _______________________________________________
>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>>>
>>>>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux