Re: cephadm trouble

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi!

Adam! Big thanx!

"ceph config rm osd.91 container_image" completly solve this trouble.
I don't understand why this happened, but at least now everything works. 

Thank you so much again!


----- Original Message -----
> From: "Fyodor Ustinov" <ufm@xxxxxx>
> To: "Adam King" <adking@xxxxxxxxxx>
> Cc: "ceph-users" <ceph-users@xxxxxxx>
> Sent: Tuesday, 1 February, 2022 18:12:16
> Subject:  Re: cephadm trouble

> Hi!
> YES! HERE IT IS!
> 
> global             basic     container_image
> quay.io/ceph/ceph@sha256:2f7f0af8663e73a422f797de605e769ae44eb0297f2a79324739404cc1765728
> *
>    osd.91         basic     container_image
>    s-8-2-1:/dev/bcache0
> 
> Two questions:
> 1. How did it get there
> 2. How to delete it - as far as I understand this field is not editable?
> 
> 
> ----- Original Message -----
>> From: "Adam King" <adking@xxxxxxxxxx>
>> To: "Fyodor Ustinov" <ufm@xxxxxx>
>> Cc: "ceph-users" <ceph-users@xxxxxxx>
>> Sent: Tuesday, 1 February, 2022 17:45:13
>> Subject: Re:  Re: cephadm trouble
> 
>> As a follow up to my previous comment, could you also post "ceph config
>> dump | grep container_image". It's related to the repo digest thing and
>> it's another way we could maybe discover where "s-8-2-1:/dev/bcache0" is
>> set as an image.
>> 
>> - Adam King
>> 
>> On Tue, Feb 1, 2022 at 8:52 AM Adam King <adking@xxxxxxxxxx> wrote:
>> 
>>> Hi Fyodor,
>>>
>>> Honestly I'm super confused by your case. Daemon add osd is meant to be a
>>> one time synchronous command so the idea that that is causing this repeated
>>> pull in this fashion is super odd. I think I would need some sort of list
>>> of commands run on this cluster or some type of reproducer. As mentioned
>>> before, cephadm definitely thinks "s-8-2-1:/dev/bcache0" is the name of a
>>> container image but I can't think of where that is set as I didn't see it
>>> in any of the posted service specs or the config options for the any of the
>>> images but it clearly must be set somewhere or we wouldn't be trying to
>>> pull that repeatedly. Never seen an issue like this before. This is a total
>>> long shot, but you could trying setting "ceph config set mgr
>>> mgr/cephadm/use_repo_digest false" and see if it at least lets you refresh
>>> the daemons and make progress (or at least gets us different things in the
>>> logs).
>>>
>>> Sorry for not being too helpful,
>>>
>>> - Adam King
>>>
>>> On Tue, Feb 1, 2022 at 3:27 AM Fyodor Ustinov <ufm@xxxxxx> wrote:
>>>
>>>> Hi!
>>>>
>>>> No mode ideas? :(
>>>>
>>>>
>>>> ----- Original Message -----
>>>> > From: "Fyodor Ustinov" <ufm@xxxxxx>
>>>> > To: "Adam King" <adking@xxxxxxxxxx>
>>>> > Cc: "ceph-users" <ceph-users@xxxxxxx>
>>>> > Sent: Friday, 28 January, 2022 23:02:26
>>>> > Subject:  Re: cephadm trouble
>>>>
>>>> > Hi!
>>>> >
>>>> >> Hmm, I'm not seeing anything that could be a cause in any of that
>>>> output. I
>>>> >> did notice, however, from your "ceph orch ls" output that none of your
>>>> >> services have been refreshed since the 24th. Cephadm typically tries to
>>>> >> refresh these things every 10 minutes so that signals something is
>>>> quite
>>>> >> wrong.
>>>> > From what I see in /var/log/ceph/cephadm.log it tries to run the same
>>>> command
>>>> > once a minute and does nothing else. That's why the status has not been
>>>> updated
>>>> > for 5 days.
>>>> >
>>>> >> Could you try running "ceph mgr fail" and if nothing seems to be
>>>> >> resolved could you post "ceph log last 200 debug cephadm". Maybe we
>>>> can see
>>>> >> if something gets stuck again after the mgr restarts.
>>>> > "ceph mgr fail" did not help.
>>>> > "ceph log last 200 debug cephadm" show again and again and again:
>>>> >
>>>> > 2022-01-28T20:57:12.792090+0000 mgr.s-26-9-24-mon-m2.nhltmq
>>>> (mgr.129738166) 349
>>>> > : cephadm [ERR] cephadm exited with an error code: 1, stderr:Pulling
>>>> container
>>>> > image s-8-2-1:/dev/bcache0...
>>>> > Non-zero exit code 125 from /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > /usr/bin/podman: stderr Error: invalid reference format
>>>> > ERROR: Failed command: /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > Traceback (most recent call last):
>>>> >  File "/usr/share/ceph/mgr/cephadm/serve.py", line 1363, in
>>>> _remote_connection
>>>> >    yield (conn, connr)
>>>> >  File "/usr/share/ceph/mgr/cephadm/serve.py", line 1256, in _run_cephadm
>>>> >    code, '\n'.join(err)))
>>>> > orchestrator._interface.OrchestratorError: cephadm exited with an error
>>>> code: 1,
>>>> > stderr:Pulling container image s-8-2-1:/dev/bcache0...
>>>> > Non-zero exit code 125 from /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > /usr/bin/podman: stderr Error: invalid reference format
>>>> > ERROR: Failed command: /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > 2022-01-28T20:58:13.092996+0000 mgr.s-26-9-24-mon-m2.nhltmq
>>>> (mgr.129738166) 392
>>>> > : cephadm [ERR] cephadm exited with an error code: 1, stderr:Pulling
>>>> container
>>>> > image s-8-2-1:/dev/bcache0...
>>>> > Non-zero exit code 125 from /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > /usr/bin/podman: stderr Error: invalid reference format
>>>> > ERROR: Failed command: /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > Traceback (most recent call last):
>>>> >  File "/usr/share/ceph/mgr/cephadm/serve.py", line 1363, in
>>>> _remote_connection
>>>> >    yield (conn, connr)
>>>> >  File "/usr/share/ceph/mgr/cephadm/serve.py", line 1256, in _run_cephadm
>>>> >    code, '\n'.join(err)))
>>>> > orchestrator._interface.OrchestratorError: cephadm exited with an error
>>>> code: 1,
>>>> > stderr:Pulling container image s-8-2-1:/dev/bcache0...
>>>> > Non-zero exit code 125 from /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> > /usr/bin/podman: stderr Error: invalid reference format
>>>> > ERROR: Failed command: /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> >
>>>> >>
>>>> >> Thanks,
>>>> >>
>>>> >> - Adam King
>>>> >>
>>>> >> On Thu, Jan 27, 2022 at 7:06 PM Fyodor Ustinov <ufm@xxxxxx> wrote:
>>>> >>
>>>> >>> Hi!
>>>> >>>
>>>> >>> I think this happened after I tried to recreate the osd with the
>>>> command
>>>> >>> "ceph orch daemon add osd s-8-2-1:/dev/bcache0"
>>>> >>>
>>>> >>>
>>>> >>> > It looks like cephadm believes "s-8-2-1:/dev/bcache0" is a container
>>>> >>> image
>>>> >>> > for some daemon. Can you provide the output of "ceph orch ls
>>>> --format
>>>> >>> > yaml",
>>>> >>>
>>>> >>> https://pastebin.com/CStBf4J0
>>>> >>>
>>>> >>> > "ceph orch upgrade status",
>>>> >>> root@s-26-9-19-mon-m1:~# ceph orch upgrade status
>>>> >>> {
>>>> >>>     "target_image": null,
>>>> >>>     "in_progress": false,
>>>> >>>     "services_complete": [],
>>>> >>>     "progress": null,
>>>> >>>     "message": ""
>>>> >>> }
>>>> >>>
>>>> >>>
>>>> >>> > "ceph config get mgr container_image",
>>>> >>> root@s-26-9-19-mon-m1:~# ceph config get mgr container_image
>>>> >>>
>>>> >>>
>>>> quay.io/ceph/ceph@sha256:2f7f0af8663e73a422f797de605e769ae44eb0297f2a79324739404cc1765728
>>>> >>>
>>>> >>>
>>>> >>> > and the values for monitoring stack container images (format is
>>>> "ceph
>>>> >>> > config get mgr mgr/cephadm/container_image_<daemon-type>" where
>>>> daemon
>>>> >>> type
>>>> >>> > is one of "prometheus", "node_exporter", "alertmanager", "grafana",
>>>> >>> > "haproxy", "keepalived").
>>>> >>> quay.io/prometheus/prometheus:v2.18.1
>>>> >>> quay.io/prometheus/node-exporter:v0.18.1
>>>> >>> quay.io/prometheus/alertmanager:v0.20.0
>>>> >>> quay.io/ceph/ceph-grafana:6.7.4
>>>> >>> docker.io/library/haproxy:2.3
>>>> >>> docker.io/arcts/keepalived
>>>> >>>
>>>> >>> >
>>>> >>> > Thanks,
>>>> >>> >
>>>> >>> > - Adam King
>>>> >>>
>>>> >>> Thanks a lot!
>>>> >>>
>>>> >>> WBR,
>>>> >>>     Fyodor.
>>>> >>>
>>>> >>> >
>>>> >>> > On Thu, Jan 27, 2022 at 9:10 AM Fyodor Ustinov <ufm@xxxxxx> wrote:
>>>> >>> >
>>>> >>> >> Hi!
>>>> >>> >>
>>>> >>> >> I rebooted the nodes with mgr and now I see the following in the
>>>> >>> >> cephadm.log:
>>>> >>> >>
>>>> >>> >> As I understand it - cephadm is trying to execute some unsuccessful
>>>> >>> >> command of mine (I wonder which one), it does not succeed, but it
>>>> keeps
>>>> >>> >> trying and trying. How do I stop it from trying?
>>>> >>> >>
>>>> >>> >> 2022-01-27 16:02:58,123 7fca7beca740 DEBUG
>>>> >>> >>
>>>> >>>
>>>> --------------------------------------------------------------------------------
>>>> >>> >> cephadm ['--image', 's-8-2-1:/dev/bcache0', 'pull']
>>>> >>> >> 2022-01-27 16:02:58,147 7fca7beca740 DEBUG /usr/bin/podman: 3.3.1
>>>> >>> >> 2022-01-27 16:02:58,249 7fca7beca740 INFO Pulling container image
>>>> >>> >> s-8-2-1:/dev/bcache0...
>>>> >>> >> 2022-01-27 16:02:58,278 7fca7beca740 DEBUG /usr/bin/podman: Error:
>>>> >>> invalid
>>>> >>> >> reference format
>>>> >>> >> 2022-01-27 16:02:58,279 7fca7beca740 INFO Non-zero exit code 125
>>>> from
>>>> >>> >> /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> >>> >> 2022-01-27 16:02:58,279 7fca7beca740 INFO /usr/bin/podman: stderr
>>>> Error:
>>>> >>> >> invalid reference format
>>>> >>> >> 2022-01-27 16:02:58,279 7fca7beca740 ERROR ERROR: Failed command:
>>>> >>> >> /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> >>> >> 2022-01-27 16:03:58,420 7f897a7a6740 DEBUG
>>>> >>> >>
>>>> >>>
>>>> --------------------------------------------------------------------------------
>>>> >>> >> cephadm ['--image', 's-8-2-1:/dev/bcache0', 'pull']
>>>> >>> >> 2022-01-27 16:03:58,443 7f897a7a6740 DEBUG /usr/bin/podman: 3.3.1
>>>> >>> >> 2022-01-27 16:03:58,547 7f897a7a6740 INFO Pulling container image
>>>> >>> >> s-8-2-1:/dev/bcache0...
>>>> >>> >> 2022-01-27 16:03:58,575 7f897a7a6740 DEBUG /usr/bin/podman: Error:
>>>> >>> invalid
>>>> >>> >> reference format
>>>> >>> >> 2022-01-27 16:03:58,577 7f897a7a6740 INFO Non-zero exit code 125
>>>> from
>>>> >>> >> /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> >>> >> 2022-01-27 16:03:58,577 7f897a7a6740 INFO /usr/bin/podman: stderr
>>>> Error:
>>>> >>> >> invalid reference format
>>>> >>> >> 2022-01-27 16:03:58,577 7f897a7a6740 ERROR ERROR: Failed command:
>>>> >>> >> /usr/bin/podman pull s-8-2-1:/dev/bcache0
>>>> >>> >>
>>>> >>> >> WBR,
>>>> >>> >>     Fyodor.
>>>> >>> >> _______________________________________________
>>>> >>> >> ceph-users mailing list -- ceph-users@xxxxxxx
>>>> >>> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>> >>> >>
>>>> >>>
>>>> > _______________________________________________
>>>> > ceph-users mailing list -- ceph-users@xxxxxxx
>>>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux