Hello and thanks for that. I am now able to retrieve logs. This is for a different OSD, one I have not been playing with trying to find out the issue. But has the same problems. -- Boot 697340cb8e134020bd4eed4a49351bc8 -- Nov 12 15:00:44 cephnode03 systemd[1]: Starting Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c... Nov 12 15:00:46 cephnode03 podman[1168]: 2021-11-12 15:00:46.418960924 +0000 GMT m=+0.338572668 container create 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:00:46 cephnode03 podman[1168]: 2021-11-12 15:00:46.641658452 +0000 GMT m=+0.561270185 container init 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:00:46 cephnode03 podman[1168]: 2021-11-12 15:00:46.716169028 +0000 GMT m=+0.635780791 container start 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:00:46 cephnode03 podman[1168]: 2021-11-12 15:00:46.716712108 +0000 GMT m=+0.636323848 container attach 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:02:44 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: start operation timed out. Terminating. Nov 12 15:02:45 cephnode03 podman[1854]: 2021-11-12 15:02:45.295203029 +0000 GMT m=+0.260481391 container create 3849fa85cedad0de442eaa41012ba0d6f7821056b66388a277b433f45134d0a4 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:02:45 cephnode03 podman[1854]: 2021-11-12 15:02:45.397836033 +0000 GMT m=+0.363114423 container init 3849fa85cedad0de442eaa41012ba0d6f7821056b66388a277b433f45134d0a4 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:02:45 cephnode03 podman[1854]: 2021-11-12 15:02:45.410880539 +0000 GMT m=+0.376158923 container start 3849fa85cedad0de442eaa41012ba0d6f7821056b66388a277b433f45134d0a4 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:02:45 cephnode03 podman[1854]: 2021-11-12 15:02:45.411070883 +0000 GMT m=+0.376349235 container attach 3849fa85cedad0de442eaa41012ba0d6f7821056b66388a277b433f45134d0a4 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:02:46 cephnode03 podman[1854]: 2021-11-12 15:02:46.317833738 +0000 GMT m=+1.283112131 container died 3849fa85cedad0de442eaa41012ba0d6f7821056b66388a277b433f45134d0a4 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:02:46 cephnode03 podman[1854]: 2021-11-12 15:02:46.400485223 +0000 GMT m=+1.365763636 container remove 3849fa85cedad0de442eaa41012ba0d6f7821056b66388a277b433f45134d0a4 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Failed with result 'timeout'. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 574 (bash) remains running after unit stopped. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 1168 (podman) remains running after unit stopped. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 1377 (conmon) remains running after unit stopped. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 1916 (conmon) remains running after unit stopped. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 1998 (podman) remains running after unit stopped. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 2023 (podman) remains running after unit stopped. Nov 12 15:02:46 cephnode03 systemd[1]: Failed to start Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:02:46 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.068s CPU time. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Scheduled restart job, restart counter is at 1. Nov 12 15:02:56 cephnode03 systemd[1]: Stopped Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.082s CPU time. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 574 (bash) in control group while starting unit. Ignoring. Nov 12 15:02:56 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 1168 (podman) in control group while starting unit. Ignoring. Nov 12 15:02:56 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 1377 (conmon) in control group while starting unit. Ignoring. Nov 12 15:02:56 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:02:56 cephnode03 systemd[1]: Starting Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c... Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 574 (bash) in control group while starting unit. Ignoring. Nov 12 15:02:56 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 1168 (podman) in control group while starting unit. Ignoring. Nov 12 15:02:56 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:02:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 1377 (conmon) in control group while starting unit. Ignoring. Nov 12 15:02:56 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:02:56 cephnode03 podman[2171]: 2021-11-12 15:02:56.897495225 +0000 GMT m=+0.271553564 container stop 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:02:56 cephnode03 podman[2171]: 2021-11-12 15:02:56.953451318 +0000 GMT m=+0.327509662 container died 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:02:57 cephnode03 podman[2171]: 2021-11-12 15:02:57.080927142 +0000 GMT m=+0.454985464 container remove 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:02:57 cephnode03 bash[2171]: 5cff11c4c70c0776658905a187c0f34ad8a8b88d399abf6f773d3c04994b58c6 Nov 12 15:02:57 cephnode03 podman[2451]: 2021-11-12 15:02:57.938588908 +0000 GMT m=+0.497876563 container create 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:02:58 cephnode03 podman[2451]: 2021-11-12 15:02:58.32025349 +0000 GMT m=+0.879541154 container init 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a785> Nov 12 15:02:58 cephnode03 podman[2451]: 2021-11-12 15:02:58.358130824 +0000 GMT m=+0.917418492 container start 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:02:58 cephnode03 podman[2451]: 2021-11-12 15:02:58.358610898 +0000 GMT m=+0.917898568 container attach 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:04:56 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: start operation timed out. Terminating. Nov 12 15:04:57 cephnode03 podman[2880]: 2021-11-12 15:04:57.172277378 +0000 GMT m=+0.192406841 container create 52129775b66c9414d5c237a3edf4a45b1d6594899b7439a65a785d93f41bf056 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:04:57 cephnode03 podman[2880]: 2021-11-12 15:04:57.282396807 +0000 GMT m=+0.302526273 container init 52129775b66c9414d5c237a3edf4a45b1d6594899b7439a65a785d93f41bf056 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:04:57 cephnode03 podman[2880]: 2021-11-12 15:04:57.307421739 +0000 GMT m=+0.327551171 container start 52129775b66c9414d5c237a3edf4a45b1d6594899b7439a65a785d93f41bf056 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:04:57 cephnode03 podman[2880]: 2021-11-12 15:04:57.307885433 +0000 GMT m=+0.328014865 container attach 52129775b66c9414d5c237a3edf4a45b1d6594899b7439a65a785d93f41bf056 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:04:58 cephnode03 podman[2880]: 2021-11-12 15:04:58.12821031 +0000 GMT m=+1.148339771 container died 52129775b66c9414d5c237a3edf4a45b1d6594899b7439a65a785d93f41bf056 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a785> Nov 12 15:04:58 cephnode03 podman[2880]: 2021-11-12 15:04:58.25156536 +0000 GMT m=+1.271694782 container remove 52129775b66c9414d5c237a3edf4a45b1d6594899b7439a65a785d93f41bf056 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Failed with result 'timeout'. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 2094 (bash) remains running after unit stopped. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 2451 (podman) remains running after unit stopped. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 2528 (conmon) remains running after unit stopped. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 2983 (conmon) remains running after unit stopped. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 3020 (podman) remains running after unit stopped. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 3061 (podman) remains running after unit stopped. Nov 12 15:04:58 cephnode03 systemd[1]: Failed to start Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:04:58 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.723s CPU time. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Scheduled restart job, restart counter is at 2. Nov 12 15:05:08 cephnode03 systemd[1]: Stopped Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.755s CPU time. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 2094 (bash) in control group while starting unit. Ignoring. Nov 12 15:05:08 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 2451 (podman) in control group while starting unit. Ignoring. Nov 12 15:05:08 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 2528 (conmon) in control group while starting unit. Ignoring. Nov 12 15:05:08 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:05:08 cephnode03 systemd[1]: Starting Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c... Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 2094 (bash) in control group while starting unit. Ignoring. Nov 12 15:05:08 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 2451 (podman) in control group while starting unit. Ignoring. Nov 12 15:05:08 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:05:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 2528 (conmon) in control group while starting unit. Ignoring. Nov 12 15:05:08 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:05:08 cephnode03 podman[3170]: 2021-11-12 15:05:08.754453546 +0000 GMT m=+0.299183916 container stop 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:05:08 cephnode03 podman[3170]: 2021-11-12 15:05:08.783446549 +0000 GMT m=+0.328176890 container died 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:05:08 cephnode03 podman[3170]: 2021-11-12 15:05:08.851135864 +0000 GMT m=+0.395866217 container remove 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:05:08 cephnode03 bash[3170]: 67775626001a63cb6127f670e8c6e0e98d368bac87b037a606008bde9ebd500c Nov 12 15:05:09 cephnode03 podman[3479]: 2021-11-12 15:05:09.80276864 +0000 GMT m=+0.386506470 container create 6dd198dfeeb73916004ce854540b26477c8f8f35f61b98c2fdbabc09541a8553 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:05:10 cephnode03 podman[3479]: 2021-11-12 15:05:10.039849134 +0000 GMT m=+0.623586958 container init 6dd198dfeeb73916004ce854540b26477c8f8f35f61b98c2fdbabc09541a8553 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:05:10 cephnode03 podman[3479]: 2021-11-12 15:05:10.16202613 +0000 GMT m=+0.745763967 container start 6dd198dfeeb73916004ce854540b26477c8f8f35f61b98c2fdbabc09541a8553 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:05:10 cephnode03 podman[3479]: 2021-11-12 15:05:10.162524955 +0000 GMT m=+0.746262780 container attach 6dd198dfeeb73916004ce854540b26477c8f8f35f61b98c2fdbabc09541a8553 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:07:08 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: start operation timed out. Terminating. Nov 12 15:07:09 cephnode03 podman[3943]: 2021-11-12 15:07:09.144930971 +0000 GMT m=+0.190925929 container create 043e7d63f534a50503de2ba07c81f0f803d10dbdba4c8b5447789e2f31326a23 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:07:09 cephnode03 podman[3943]: 2021-11-12 15:07:09.254707415 +0000 GMT m=+0.300702370 container init 043e7d63f534a50503de2ba07c81f0f803d10dbdba4c8b5447789e2f31326a23 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:07:09 cephnode03 podman[3943]: 2021-11-12 15:07:09.295296356 +0000 GMT m=+0.341291325 container start 043e7d63f534a50503de2ba07c81f0f803d10dbdba4c8b5447789e2f31326a23 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:07:09 cephnode03 podman[3943]: 2021-11-12 15:07:09.295479945 +0000 GMT m=+0.341474869 container attach 043e7d63f534a50503de2ba07c81f0f803d10dbdba4c8b5447789e2f31326a23 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:07:10 cephnode03 podman[3943]: 2021-11-12 15:07:10.105620642 +0000 GMT m=+1.151615597 container died 043e7d63f534a50503de2ba07c81f0f803d10dbdba4c8b5447789e2f31326a23 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:07:10 cephnode03 podman[3943]: 2021-11-12 15:07:10.329322681 +0000 GMT m=+1.375317649 container remove 043e7d63f534a50503de2ba07c81f0f803d10dbdba4c8b5447789e2f31326a23 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Failed with result 'timeout'. Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 3111 (bash) remains running after unit stopped. Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 3479 (podman) remains running after unit stopped. Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 3581 (conmon) remains running after unit stopped. Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 4041 (conmon) remains running after unit stopped. Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 4084 (podman) remains running after unit stopped. Nov 12 15:07:10 cephnode03 systemd[1]: Failed to start Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:07:10 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.540s CPU time. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Scheduled restart job, restart counter is at 3. Nov 12 15:07:20 cephnode03 systemd[1]: Stopped Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.572s CPU time. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 3111 (bash) in control group while starting unit. Ignoring. Nov 12 15:07:20 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 3479 (podman) in control group while starting unit. Ignoring. Nov 12 15:07:20 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 3581 (conmon) in control group while starting unit. Ignoring. Nov 12 15:07:20 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:07:20 cephnode03 systemd[1]: Starting Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c... Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 3111 (bash) in control group while starting unit. Ignoring. Nov 12 15:07:20 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 3479 (podman) in control group while starting unit. Ignoring. Nov 12 15:07:20 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:07:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 3581 (conmon) in control group while starting unit. Ignoring. Nov 12 15:07:20 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:07:20 cephnode03 podman[4312]: 2021-11-12 15:07:20.935350131 +0000 GMT m=+0.151259617 container remove 6dd198dfeeb73916004ce854540b26477c8f8f35f61b98c2fdbabc09541a8553 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:07:21 cephnode03 podman[4504]: 2021-11-12 15:07:21.813813737 +0000 GMT m=+0.344842861 container create 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:07:21 cephnode03 bash[3479]: time="2021-11-12T15:07:21Z" level=error msg="Cannot get exit code: died not found: unable to find event" Nov 12 15:07:21 cephnode03 podman[4504]: 2021-11-12 15:07:21.957863052 +0000 GMT m=+0.488892169 container init 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:07:21 cephnode03 podman[4504]: 2021-11-12 15:07:21.986447108 +0000 GMT m=+0.517476234 container start 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:07:21 cephnode03 podman[4504]: 2021-11-12 15:07:21.986574131 +0000 GMT m=+0.517603271 container attach 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:09:20 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: start operation timed out. Terminating. Nov 12 15:09:21 cephnode03 podman[5007]: 2021-11-12 15:09:21.481588083 +0000 GMT m=+0.222154705 container create 9ec0bc0e740fd527861edec9b831e3b1bcb5318c8b12530c2d840fd0b2d0ac6b (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:09:21 cephnode03 podman[5007]: 2021-11-12 15:09:21.657495698 +0000 GMT m=+0.398062308 container init 9ec0bc0e740fd527861edec9b831e3b1bcb5318c8b12530c2d840fd0b2d0ac6b (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:09:21 cephnode03 podman[5007]: 2021-11-12 15:09:21.669402917 +0000 GMT m=+0.409969527 container start 9ec0bc0e740fd527861edec9b831e3b1bcb5318c8b12530c2d840fd0b2d0ac6b (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:09:21 cephnode03 podman[5007]: 2021-11-12 15:09:21.669590773 +0000 GMT m=+0.410157368 container attach 9ec0bc0e740fd527861edec9b831e3b1bcb5318c8b12530c2d840fd0b2d0ac6b (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:09:22 cephnode03 podman[5007]: 2021-11-12 15:09:22.532873483 +0000 GMT m=+1.273440122 container died 9ec0bc0e740fd527861edec9b831e3b1bcb5318c8b12530c2d840fd0b2d0ac6b (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:09:22 cephnode03 podman[5007]: 2021-11-12 15:09:22.664820144 +0000 GMT m=+1.405386768 container remove 9ec0bc0e740fd527861edec9b831e3b1bcb5318c8b12530c2d840fd0b2d0ac6b (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a> Nov 12 15:09:22 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Failed with result 'timeout'. Nov 12 15:09:22 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 4173 (bash) remains running after unit stopped. Nov 12 15:09:22 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 4504 (podman) remains running after unit stopped. Nov 12 15:09:22 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Unit process 4622 (conmon) remains running after unit stopped. Nov 12 15:09:22 cephnode03 systemd[1]: Failed to start Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:09:22 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.560s CPU time. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Scheduled restart job, restart counter is at 4. Nov 12 15:09:32 cephnode03 systemd[1]: Stopped Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Consumed 1.567s CPU time. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 4173 (bash) in control group while starting unit. Ignoring. Nov 12 15:09:32 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 4504 (podman) in control group while starting unit. Ignoring. Nov 12 15:09:32 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 4622 (conmon) in control group while starting unit. Ignoring. Nov 12 15:09:32 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:09:32 cephnode03 systemd[1]: Starting Ceph osd.8 for d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c... Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 4173 (bash) in control group while starting unit. Ignoring. Nov 12 15:09:32 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 4504 (podman) in control group while starting unit. Ignoring. Nov 12 15:09:32 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:09:32 cephnode03 systemd[1]: ceph-d82ced8c-1f6e-11ec- a2e4-00fd45fcaf9c@osd.8.service: Found left-over process 4622 (conmon) in control group while starting unit. Ignoring. Nov 12 15:09:32 cephnode03 systemd[1]: This usually indicates unclean termination of a previous run, or service implementation deficiencies. Nov 12 15:09:33 cephnode03 podman[5319]: 2021-11-12 15:09:33.046838444 +0000 GMT m=+0.214927733 container stop 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:09:33 cephnode03 podman[5319]: 2021-11-12 15:09:33.085449517 +0000 GMT m=+0.253538787 container died 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:09:33 cephnode03 podman[5319]: 2021-11-12 15:09:33.23814693 +0000 GMT m=+0.406236223 container remove 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:09:33 cephnode03 bash[5319]: 3ccc8458beb0820484c501621bec99011c5af24744fdfebd47b11f342a47fb72 Nov 12 15:09:34 cephnode03 podman[5606]: 2021-11-12 15:09:34.14523469 +0000 GMT m=+0.442976747 container create 30d3afe7236ae8f5553ed3ad108596b22a74ab17fb132ccf433caba7ca959de2 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> Nov 12 15:09:34 cephnode03 podman[5606]: 2021-11-12 15:09:34.329560878 +0000 GMT m=+0.627302934 container init 30d3afe7236ae8f5553ed3ad108596b22a74ab17fb132ccf433caba7ca959de2 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a78> Nov 12 15:09:34 cephnode03 podman[5606]: 2021-11-12 15:09:34.422080643 +0000 GMT m=+0.719822670 container start 30d3afe7236ae8f5553ed3ad108596b22a74ab17fb132ccf433caba7ca959de2 (image=quay.io/ceph/ ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a7> I am trying to understand what is happening. Any suggestions are most welcome. Thanks, Stephen. On Saturday, 13 November 2021 14:38:28 GMT 胡 玮文 wrote: > Hi Stephen, > > I think this output you posted is pretty normal, there is no systemd in the > container, thus the error. > You still need to find the logs. You may try “sudo cephadm logs --name > osd.0”. If that still fails, you should try to run the ceph-osd daemon > manually. > Weiwen Hu > > 发件人: Stephen J. Thompson<mailto:stephen@xxxxxxxxxxxxxxxxxxxxx> > 发送时间: 2021年11月13日 20:19 > 收件人: ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> > 抄送: Stephen J. Thompson<mailto:stephen@xxxxxxxxxxxxxxxxxxxxx> > 主题: Re: OSDs not starting up > > Hello all, > > I am still digging into this. I disabled this osd from starting at boot, > then rebooted the node. > > I then tried doing the following: > > sudo cephadm shell > > And the following was the result. To me it seems to indicate that the OSD > drive is ok and can be decrypted but does still not run. > > Inferring fsid d82ced8c-1f6e-11ec-a2e4-00fd45fcaf9c > Using recent ceph image quay.io/ceph/ > ceph@sha256:285d2abb6e74bdc6e15e1af585aa19f132045651b7a80eb77b9cec8a785ff330 > root@cephnode02:/# ceph-volume lvm activate 0 > 2f5d17d3-3308-4f42-867c-7ec8639bde18 > Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 > Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-0/ > lockbox.keyring --create-keyring --name client.osd-lockbox. > 2f5d17d3-3308-4f42-867c-7ec8639bde18 --add-key > AQDNzFFh11ooIRAAru8xYXAXL1/n/ 75AOJm2KA== > stdout: creating /var/lib/ceph/osd/ceph-0/lockbox.keyring > added entity client.osd-lockbox.2f5d17d3-3308-4f42-867c-7ec8639bde18 > auth(key=AQDNzFFh11ooIRAAru8xYXAXL1/n/75AOJm2KA==) > Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ > lockbox.keyring > Running command: /usr/bin/ceph --cluster ceph --name client.osd-lockbox. > 2f5d17d3-3308-4f42-867c-7ec8639bde18 --keyring /var/lib/ceph/osd/ceph-0/ > lockbox.keyring config-key get dm-crypt/osd/ > 2f5d17d3-3308-4f42-867c-7ec8639bde18/luks > Running command: /usr/sbin/cryptsetup --key-file - --allow-discards luksOpen > / dev/ceph-779b00ee-dbb5-4542-859c-b6338b01efe0/osd- > block-2f5d17d3-3308-4f42-867c-7ec8639bde18 6oP2t6-fuMx-25XU-KiIt-A28Q-fzng- > K7ouf6 > Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 > Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir > -- dev /dev/mapper/6oP2t6-fuMx-25XU-KiIt-A28Q-fzng-K7ouf6 --path > /var/lib/ceph/ osd/ceph-0 --no-mon-config > Running command: /usr/bin/ln -snf > /dev/mapper/6oP2t6-fuMx-25XU-KiIt-A28Q-fzng- K7ouf6 > /var/lib/ceph/osd/ceph-0/block > Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-0/block > Running command: /usr/bin/chown -R ceph:ceph /dev/dm-7 > Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 > Running command: /usr/bin/systemctl enable ceph- > volume@lvm-0-2f5d17d3-3308-4f42-867c-7ec8639bde18 > stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph- > volume@lvm-0-2f5d17d3-3308-4f42-867c-7ec8639bde18.service -> > /usr/lib/systemd/ system/ceph-volume@.service. > Running command: /usr/bin/systemctl enable --runtime ceph-osd@0 > stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph- > osd@0.service -> /usr/lib/systemd/system/ceph-osd@.service. > Running command: /usr/bin/systemctl start ceph-osd@0 > stderr: Failed to connect to bus: No such file or directory > --> RuntimeError: command returned non-zero exit status: 1 > > Any ideas? > > Thanks, > > Stephen > > > > On Friday, 12 November 2021 21:32:41 GMT Stephen J. Thompson wrote: > > > Hello Igor, > > > > > > > > The OSD logs a empty. > > > > > > > > Enclosed is the end of the ceph-volume.log > > > > > > > > Regards, > > > > > > > > Stephen > > > > > > > > [2021-11-12 21:23:54,191][ceph_volume.main][INFO ] Running command: > > ceph- > > volume inventory --format=json --filter-for-batch > > [2021-11-12 21:23:54,194][ceph_volume.process][INFO ] Running command: > > /usr/ bin/lsblk -plno KNAME,NAME,TYPE > > [2021-11-12 21:23:54,202][ceph_volume.process][INFO ] stdout /dev/sda > > /dev/ sda > > disk > > [2021-11-12 21:23:54,202][ceph_volume.process][INFO ] stdout /dev/sdb > > /dev/ sdb > > disk > > [2021-11-12 21:23:54,202][ceph_volume.process][INFO ] stdout /dev/sdc > > /dev/ sdc > > disk > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/sdd > > /dev/ sdd > > disk > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/sdd1 > > /dev/ sdd1 > > part > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/sdd2 > > /dev/ sdd2 > > part > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/sdd5 > > /dev/ sdd5 > > part > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-0 > > /dev/ mapper/ceph--779b00ee--dbb5--4542--859c--b6338b01efe0-osd-- > > block--2f5d17d3--3308--4f42--867c--7ec8639bde18 lvm > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-1 > > /dev/ mapper/ceph--41cbaf58--d703--4bae--9095--f1b590b8337b-osd-- > > block--99e1e1b2--3cfe--48e0--8627--acfe5008c1c5 lvm > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-2 > > /dev/ mapper/ceph--09c84540--06e3--496f--bf90--45f59748768a-osd-- > > block--60ad9fb3--8144--4d06--83e6--b90e9d52dcaa lvm > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-3 > > /dev/ mapper/sda5_crypt > > crypt > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-4 > > /dev/ mapper/cephnode02--vg-root > > lvm > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-5 > > /dev/ mapper/cephnode02--vg-swap_1 > > lvm > > [2021-11-12 21:23:54,203][ceph_volume.process][INFO ] stdout /dev/dm-6 > > /dev/ mapper/fwR6Nz-3DbF-Ac2K-jhVp-DtsV-A1mt-VMUefw > > crypt > > [2021-11-12 21:23:54,210][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sda -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:54,910][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sda > > [2021-11-12 21:23:54,917][ceph_volume.process][INFO ] stdout NAME="sda" > > KNAME="sda" MAJ:MIN="8:0" FSTYPE="LVM2_member" MOUNTPOINT="" LABEL="" > > UUID="G9naIi-2sgq-FznV-3a5U-8VH6-0G9C-HRQvGf" RO="0" RM="0" > > MODEL="TOSHIBA > > HDWN180 " SIZE="7.3T" STATE="running" OWNER="root" GROUP="disk" > > MODE="brw- > > rw----" ALIGNMENT="0" PHY-SEC="4096" LOG-SEC="512" ROTA="1" SCHED="mq- > > deadline" TYPE="disk" DISC-ALN="0" DISC-GRAN="0B" DISC-MAX="0B" > > DISC-ZERO="0" PKNAME="" PARTLABEL="" > > [2021-11-12 21:23:54,918][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sda > > [2021-11-12 21:23:54,922][ceph_volume.process][INFO ] stdout /dev/sda: > > UUID="G9naIi-2sgq-FznV-3a5U-8VH6-0G9C-HRQvGf" VERSION="LVM2 001" > > TYPE="LVM2_member" USAGE="raid" > > [2021-11-12 21:23:54,923][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sda > > [2021-11-12 21:23:54,978][ceph_volume.process][INFO ] stdout > > ceph-779b00ee- > > dbb5-4542-859c-b6338b01efe0";"1";"1";"wz--n-";"1907721";"0";"4194304 > > [2021-11-12 21:23:54,979][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -o lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size /dev/sda > > [2021-11-12 21:23:55,118][ceph_volume.process][INFO ] stdout > > ceph.block_device=/dev/ceph-779b00ee-dbb5-4542-859c-b6338b01efe0/osd- > > block-2f5d17d3-3308-4f42-867c-7ec8639bde18,ceph.block_uuid=6oP2t6-fuMx-25X > > U- > > KiIt-A28Q-fzng-K7ouf6,ceph.cephx_lockbox_secret=AQDNzFFh11ooIRAAru8xYXAXL > > 1/ n/ 75AOJm2KA==,ceph.cluster_fsid=d82ced8c-1f6e-11ec- > > a2e4-00fd45fcaf9c,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph > > .e > > ncrypted=1,ceph.osd_fsid=2f5d17d3-3308-4f42-867c-7ec8639bde18,ceph.osd_id > > =0, > > ceph.osdspec_affinity=default_drive_group,ceph.type=block,ceph.vdo=0";"/ > > dev/ceph-779b00ee-dbb5-4542-859c-b6338b01efe0/osd- > > block-2f5d17d3-3308-4f42-867c-7ec8639bde18";"osd- > > block-2f5d17d3-3308-4f42-867c-7ec8639bde18";"ceph-779b00ee-dbb5-4542-859c- > > > > b6338b01efe0";"6oP2t6-fuMx-25XU-KiIt-A28Q-fzng-K7ouf6";"8001561821184 > > [2021-11-12 21:23:55,119][ceph_volume.util.disk][INFO ] opening device > > /dev/ sda to check for BlueStore label > > [2021-11-12 21:23:55,119][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sda > > [2021-11-12 21:23:55,131][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-path/pci-0000:00:1f.2-ata-1.0 > > /dev/disk/by-id/wwn-0x500003998c8012f6 / > > dev/disk/by-id/ata-TOSHIBA_HDWN180_99RZK0PIFAVG /dev/disk/by-path/ > > pci-0000:00:1f.2-ata-1 /dev/disk/by-id/lvm-pv-uuid-G9naIi-2sgq- > > FznV-3a5U-8VH6-0G9C-HRQvGf > > [2021-11-12 21:23:55,131][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/sda [2021-11-12 21:23:55,131][ceph_volume.process][INFO ] > > stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata1/host0/target0:0:0/0:0:0:0/block/sda > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > DEVTYPE=disk > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=128 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:55,132][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=66362 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=66362 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=7200 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:55,133][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=LVM2_member > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_FS_USAGE=raid [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] > > stdout > > ID_FS_UUID=G9naIi-2sgq-FznV-3a5U-8VH6-0G9C-HRQvGf > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=G9naIi-2sgq-FznV-3a5U-8VH6-0G9C-HRQvGf > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=LVM2 001 > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_MODEL=TOSHIBA_HDWN180 > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=TOSHIBA\x20HDWN180\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-1.0 > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-1 > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-1_0 > > [2021-11-12 21:23:55,134][ceph_volume.process][INFO ] stdout > > ID_REVISION=GX2M [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] > > stdout > > ID_SERIAL=TOSHIBA_HDWN180_99RZK0PIFAVG > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=99RZK0PIFAVG > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > ID_WWN=0x500003998c8012f6 > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > ID_WWN_WITH_EXTENSION=0x500003998c8012f6 > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout MINOR=0 > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] > > stdout SYSTEMD_ALIAS=/ dev/block/8:0 > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > SYSTEMD_READY=1 [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] > > stdout > > SYSTEMD_WANTS=lvm2-pvscan@8:0.service > > [2021-11-12 21:23:55,135][ceph_volume.process][INFO ] stdout > > TAGS=:systemd: [2021-11-12 21:23:55,136][ceph_volume.process][INFO ] > > stdout > > USEC_INITIALIZED=4467643 > > [2021-11-12 21:23:55,136][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdb -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:55,198][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdb > > [2021-11-12 21:23:55,205][ceph_volume.process][INFO ] stdout NAME="sdb" > > KNAME="sdb" MAJ:MIN="8:16" FSTYPE="LVM2_member" MOUNTPOINT="" LABEL="" > > UUID="SOElW1-Wjj1-TV8Q-TKsa-eRBG-gzGr-aCsNdu" RO="0" RM="0" MODEL="WDC > > WD30EFRX-68E" SIZE="2.7T" STATE="running" OWNER="root" GROUP="disk" > > MODE="brw- rw----" ALIGNMENT="0" PHY-SEC="4096" LOG-SEC="512" ROTA="1" > > SCHED="mq- deadline" TYPE="disk" DISC-ALN="0" DISC-GRAN="0B" > > DISC-MAX="0B" > > DISC-ZERO="0" PKNAME="" PARTLABEL="" > > [2021-11-12 21:23:55,205][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdb > > [2021-11-12 21:23:55,209][ceph_volume.process][INFO ] stdout /dev/sdb: > > UUID="SOElW1-Wjj1-TV8Q-TKsa-eRBG-gzGr-aCsNdu" VERSION="LVM2 001" > > TYPE="LVM2_member" USAGE="raid" > > [2021-11-12 21:23:55,209][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdb > > [2021-11-12 21:23:55,266][ceph_volume.process][INFO ] stdout > > ceph-41cbaf58- > > d703-4bae-9095-f1b590b8337b";"1";"1";"wz--n-";"715396";"0";"4194304 > > [2021-11-12 21:23:55,267][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -o lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size /dev/sdb > > [2021-11-12 21:23:55,338][ceph_volume.process][INFO ] stdout > > ceph.block_device=/dev/ceph-41cbaf58-d703-4bae-9095-f1b590b8337b/osd- > > block-99e1e1b2-3cfe-48e0-8627-acfe5008c1c5,ceph.block_uuid=fwR6Nz-3DbF-Ac2 > > K- > > jhVp-DtsV-A1mt-VMUefw,ceph.cephx_lockbox_secret=AQAyQ45hPpzqDxAAx2vyba/ > > tTeQmZl1Sllh8Mg==,ceph.cluster_fsid=d82ced8c-1f6e-11ec- > > a2e4-00fd45fcaf9c,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph > > .e ncrypted=1,ceph.osd_fsid=99e1e1b2-3cfe-48e0-8627- > > acfe5008c1c5,ceph.osd_id=13,ceph.osdspec_affinity=default_drive_group,ceph > > . type=block,ceph.vdo=0";"/ > > dev/ceph-41cbaf58-d703-4bae-9095-f1b590b8337b/osd- > > block-99e1e1b2-3cfe-48e0-8627-acfe5008c1c5";"osd- > > block-99e1e1b2-3cfe-48e0-8627-acfe5008c1c5";"ceph-41cbaf58-d703-4bae-9095- > > > > f1b590b8337b";"fwR6Nz-3DbF-Ac2K-jhVp-DtsV-A1mt-VMUefw";"3000588304384 > > [2021-11-12 21:23:55,339][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdb to check for BlueStore label > > [2021-11-12 21:23:55,339][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdb > > [2021-11-12 21:23:55,351][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ > > disk/by-id/lvm-pv-uuid-SOElW1-Wjj1-TV8Q-TKsa-eRBG-gzGr-aCsNdu > > /dev/disk/by-id/ ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N99Z68Y4 > > /dev/disk/by-id/ > > wwn-0x50014ee20a61fcc0 /dev/disk/by-path/pci-0000:00:1f.2-ata-3 > > /dev/disk/by- path/pci-0000:00:1f.2-ata-3.0 > > [2021-11-12 21:23:55,351][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/sdb [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] > > stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata3/host2/target2:0:0/2:0:0:0/block/sdb > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > DEVTYPE=disk > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PUIS=1 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PUIS_ENABLED=0 > > [2021-11-12 21:23:55,352][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=420 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=420 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=5400 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:55,353][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=LVM2_member > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_FS_USAGE=raid [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] > > stdout > > ID_FS_UUID=SOElW1-Wjj1-TV8Q-TKsa-eRBG-gzGr-aCsNdu > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=SOElW1-Wjj1-TV8Q-TKsa-eRBG-gzGr-aCsNdu > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=LVM2 001 > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_MODEL=WDC_WD30EFRX-68EUZN0 > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=WDC\x20WD30EFRX-68EUZN0\x20\x20\x20\x20\x20\x20\x20\x20\x20\x > > 20 \x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-3.0 > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-3 > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-3_0 > > [2021-11-12 21:23:55,354][ceph_volume.process][INFO ] stdout > > ID_REVISION=80.00A80 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > ID_SERIAL=WDC_WD30EFRX-68EUZN0_WD-WCC4N99Z68Y4 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=WD-WCC4N99Z68Y4 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > ID_WWN=0x50014ee20a61fcc0 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > ID_WWN_WITH_EXTENSION=0x50014ee20a61fcc0 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout MINOR=16 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] > > stdout SYSTEMD_ALIAS=/ dev/block/8:16 > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > SYSTEMD_READY=1 [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] > > stdout > > SYSTEMD_WANTS=lvm2-pvscan@8:16.service > > [2021-11-12 21:23:55,355][ceph_volume.process][INFO ] stdout > > TAGS=:systemd: [2021-11-12 21:23:55,356][ceph_volume.process][INFO ] > > stdout > > USEC_INITIALIZED=4535845 > > [2021-11-12 21:23:55,356][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdc -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:55,406][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdc > > [2021-11-12 21:23:55,412][ceph_volume.process][INFO ] stdout NAME="sdc" > > KNAME="sdc" MAJ:MIN="8:32" FSTYPE="LVM2_member" MOUNTPOINT="" LABEL="" > > UUID="ywCsJt-kk1E-rTKp-eBLr-ffSo-IF0j-3ntdjZ" RO="0" RM="0" > > MODEL="TOSHIBA > > HDWD130 " SIZE="2.7T" STATE="running" OWNER="root" GROUP="disk" > > MODE="brw- > > rw----" ALIGNMENT="0" PHY-SEC="4096" LOG-SEC="512" ROTA="1" SCHED="mq- > > deadline" TYPE="disk" DISC-ALN="0" DISC-GRAN="0B" DISC-MAX="0B" > > DISC-ZERO="0" PKNAME="" PARTLABEL="" > > [2021-11-12 21:23:55,413][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdc > > [2021-11-12 21:23:55,416][ceph_volume.process][INFO ] stdout /dev/sdc: > > UUID="ywCsJt-kk1E-rTKp-eBLr-ffSo-IF0j-3ntdjZ" VERSION="LVM2 001" > > TYPE="LVM2_member" USAGE="raid" > > [2021-11-12 21:23:55,417][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdc > > [2021-11-12 21:23:55,466][ceph_volume.process][INFO ] stdout > > ceph-09c84540-06e3-496f-bf90-45f59748768a";"1";"1";"wz-- > > n-";"715396";"0";"4194304 > > [2021-11-12 21:23:55,466][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -o lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size /dev/sdc > > [2021-11-12 21:23:55,542][ceph_volume.process][INFO ] stdout > > ceph.block_device=/dev/ceph-09c84540-06e3-496f-bf90-45f59748768a/osd- > > block-60ad9fb3-8144-4d06-83e6-b90e9d52dcaa,ceph.block_uuid=h352xv-O3aB-qRA > > Q- rWUa-alhz-7b9H-o30rfs,ceph.cephx_lockbox_secret=AQB+/ > > FFhtwkAIxAAWVWk87T5duTfEIDtQGGqjg==,ceph.cluster_fsid=d82ced8c-1f6e-11ec- > > a2e4-00fd45fcaf9c,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph > > .e ncrypted=1,ceph.osd_fsid=60ad9fb3-8144-4d06-83e6- > > b90e9d52dcaa,ceph.osd_id=4,ceph.osdspec_affinity=default_drive_group,ceph. > > t ype=block,ceph.vdo=0";"/ > > dev/ceph-09c84540-06e3-496f-bf90-45f59748768a/osd- > > block-60ad9fb3-8144-4d06-83e6-b90e9d52dcaa";"osd- > > block-60ad9fb3-8144-4d06-83e6-b90e9d52dcaa";"ceph-09c84540-06e3-496f- > > bf90-45f59748768a";"h352xv-O3aB-qRAQ-rWUa-alhz-7b9H-o30rfs";"3000588304384 > > > > [2021-11-12 21:23:55,543][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdc to check for BlueStore label > > [2021-11-12 21:23:55,543][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdc > > [2021-11-12 21:23:55,555][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ > > disk/by-id/lvm-pv-uuid-ywCsJt-kk1E-rTKp-eBLr-ffSo-IF0j-3ntdjZ > > /dev/disk/by-id/ wwn-0x5000039fe6d5d3d9 > > /dev/disk/by-path/pci-0000:00:1f.2-ata-4.0 /dev/disk/ > > by-path/pci-0000:00:1f.2-ata-4 > > /dev/disk/by-id/ata-TOSHIBA_HDWD130_674K0JUAS [2021-11-12 > > 21:23:55,556][ceph_volume.process][INFO ] stdout DEVNAME=/dev/sdc > > [2021-11-12 21:23:55,556][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata4/host3/target3:0:0/3:0:0:0/block/sdc > > [2021-11-12 21:23:55,556][ceph_volume.process][INFO ] stdout > > DEVTYPE=disk > > [2021-11-12 21:23:55,556][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:55,556][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:55,556][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:55,556][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=0 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PUIS=1 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PUIS_ENABLED=0 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:55,557][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=492 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=7200 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=LVM2_member > > [2021-11-12 21:23:55,558][ceph_volume.process][INFO ] stdout > > ID_FS_USAGE=raid [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] > > stdout > > ID_FS_UUID=ywCsJt-kk1E-rTKp-eBLr-ffSo-IF0j-3ntdjZ > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=ywCsJt-kk1E-rTKp-eBLr-ffSo-IF0j-3ntdjZ > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=LVM2 001 > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_MODEL=TOSHIBA_HDWD130 > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=TOSHIBA\x20HDWD130\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-4.0 > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-4 > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-4_0 > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_REVISION=MX6OACF0 > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_SERIAL=TOSHIBA_HDWD130_674K0JUAS > > [2021-11-12 21:23:55,559][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=674K0JUAS > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout > > ID_WWN=0x5000039fe6d5d3d9 > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout > > ID_WWN_WITH_EXTENSION=0x5000039fe6d5d3d9 > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout MINOR=32 > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] > > stdout SYSTEMD_ALIAS=/ dev/block/8:32 > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout > > SYSTEMD_READY=1 [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] > > stdout > > SYSTEMD_WANTS=lvm2-pvscan@8:32.service > > [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] stdout > > TAGS=:systemd: [2021-11-12 21:23:55,560][ceph_volume.process][INFO ] > > stdout > > USEC_INITIALIZED=4505597 > > [2021-11-12 21:23:55,561][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:55,622][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd > > [2021-11-12 21:23:55,630][ceph_volume.process][INFO ] stdout NAME="sdd" > > KNAME="sdd" MAJ:MIN="8:48" FSTYPE="" MOUNTPOINT="" LABEL="" UUID="" > > RO="0" > > RM="0" MODEL="CT240BX500SSD1 " SIZE="223.6G" STATE="running" > > OWNER="root" > > GROUP="disk" MODE="brw-rw----" ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" > > ROTA="0" SCHED="mq-deadline" TYPE="disk" DISC-ALN="0" DISC-GRAN="512B" > > DISC- MAX="2G" DISC-ZERO="0" PKNAME="" PARTLABEL="" > > [2021-11-12 21:23:55,631][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdd > > [2021-11-12 21:23:55,636][ceph_volume.process][INFO ] stdout /dev/sdd: > > PTUUID="9a4155cb" PTTYPE="dos" > > [2021-11-12 21:23:55,636][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd > > [2021-11-12 21:23:55,689][ceph_volume.process][INFO ] stderr Cannot use > > /dev/ sdd: device is partitioned > > [2021-11-12 21:23:55,690][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd2 > > [2021-11-12 21:23:55,750][ceph_volume.process][INFO ] stderr Cannot use > > /dev/ sdd2: device is too small (pv_min_size) > > [2021-11-12 21:23:55,750][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd5 > > [2021-11-12 21:23:55,818][ceph_volume.process][INFO ] stderr Failed to > > find physical volume "/dev/sdd5". > > [2021-11-12 21:23:55,818][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd1 > > [2021-11-12 21:23:55,882][ceph_volume.process][INFO ] stderr Failed to > > find physical volume "/dev/sdd1". > > [2021-11-12 21:23:55,883][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd2 -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:55,958][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd2 > > [2021-11-12 21:23:55,966][ceph_volume.process][INFO ] stdout NAME="sdd2" > > KNAME="sdd2" MAJ:MIN="8:50" FSTYPE="" MOUNTPOINT="" LABEL="" UUID="" > > RO="0" RM="0" MODEL="" SIZE="1K" STATE="" OWNER="root" GROUP="disk" > > MODE="brw-rw----" ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" ROTA="0" > > SCHED="mq-deadline" TYPE="part" DISC-ALN="0" DISC-GRAN="512B" > > DISC-MAX="2G" DISC-ZERO="0" PKNAME="sdd" PARTLABEL="" > > [2021-11-12 21:23:55,967][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdd2 > > [2021-11-12 21:23:55,971][ceph_volume.process][INFO ] stdout /dev/sdd2: > > PTUUID="4808244c" PTTYPE="dos" PART_ENTRY_SCHEME="dos" > > PART_ENTRY_UUID="9a4155cb-02" PART_ENTRY_TYPE="0x5" PART_ENTRY_NUMBER="2" > > PART_ENTRY_OFFSET="1001470" PART_ENTRY_SIZE="467859458" > > PART_ENTRY_DISK="8:48" [2021-11-12 > > 21:23:55,973][ceph_volume.process][INFO > > ] Running command: /usr/ sbin/pvs --noheadings --readonly --units=b > > --nosuffix --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd2 > > [2021-11-12 21:23:56,029][ceph_volume.process][INFO ] stderr Cannot use > > /dev/ sdd2: device is too small (pv_min_size) > > [2021-11-12 21:23:56,030][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd2 to check for BlueStore label > > [2021-11-12 21:23:56,031][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,031][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd2 to check for BlueStore label > > [2021-11-12 21:23:56,031][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,032][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd2 > > [2021-11-12 21:23:56,042][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-path/pci-0000:00:1f.2-ata-5.0-part2 > > /dev/disk/by-id/ata- > > CT240BX500SSD1_1944E3D4E7BB-part2 /dev/disk/by-partuuid/9a4155cb-02 > > /dev/disk/ by-path/pci-0000:00:1f.2-ata-5-part2 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/ sdd2 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd/s > > dd 2 [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > DEVTYPE=partition > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,043][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,044][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,044][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,044][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,044][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,045][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_DISK=8:48 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_NUMBER=2 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_OFFSET=1001470 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SCHEME=dos > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SIZE=467859458 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_TYPE=0x5 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_UUID=9a4155cb-02 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=4808244c > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,046][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout MINOR=50 > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout PARTN=2 > > [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,047][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,047][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4428059 > > [2021-11-12 21:23:56,048][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd5 -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:56,118][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd5 > > [2021-11-12 21:23:56,126][ceph_volume.process][INFO ] stdout NAME="sdd5" > > KNAME="sdd5" MAJ:MIN="8:53" FSTYPE="crypto_LUKS" MOUNTPOINT="" LABEL="" > > UUID="fbfc2e93-1c31-469b-80ce-0805c065be6f" RO="0" RM="0" MODEL="" > > SIZE="223.1G" STATE="" OWNER="root" GROUP="disk" MODE="brw-rw----" > > ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" ROTA="0" SCHED="mq-deadline" > > TYPE="part" DISC-ALN="0" DISC-GRAN="512B" DISC-MAX="2G" DISC-ZERO="0" > > PKNAME="sdd" PARTLABEL="" > > [2021-11-12 21:23:56,127][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdd5 > > [2021-11-12 21:23:56,131][ceph_volume.process][INFO ] stdout /dev/sdd5: > > VERSION="2" UUID="fbfc2e93-1c31-469b-80ce-0805c065be6f" > > TYPE="crypto_LUKS" > > USAGE="crypto" PART_ENTRY_SCHEME="dos" PART_ENTRY_UUID="9a4155cb-05" > > PART_ENTRY_TYPE="0x83" PART_ENTRY_NUMBER="5" PART_ENTRY_OFFSET="1001472" > > PART_ENTRY_SIZE="467859456" PART_ENTRY_DISK="8:48" > > [2021-11-12 21:23:56,132][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd5 > > [2021-11-12 21:23:56,198][ceph_volume.process][INFO ] stderr Failed to > > find physical volume "/dev/sdd5". > > [2021-11-12 21:23:56,198][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd5 to check for BlueStore label > > [2021-11-12 21:23:56,199][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,199][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd5 to check for BlueStore label > > [2021-11-12 21:23:56,199][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,199][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd5 > > [2021-11-12 21:23:56,212][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-uuid/fbfc2e93-1c31-469b-80ce-0805c065be6f > > /dev/disk/by-partuuid/ 9a4155cb-05 > > /dev/disk/by-path/pci-0000:00:1f.2-ata-5.0-part5 /dev/disk/by- > > path/pci-0000:00:1f.2-ata-5-part5 /dev/disk/by-id/ata- > > CT240BX500SSD1_1944E3D4E7BB-part5 > > [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/ sdd5 > > [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd/s > > dd 5 [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout > > DEVTYPE=partition > > [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,213][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,214][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=crypto_LUKS > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_FS_USAGE=crypto > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_FS_UUID=fbfc2e93-1c31-469b-80ce-0805c065be6f > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=fbfc2e93-1c31-469b-80ce-0805c065be6f > > [2021-11-12 21:23:56,215][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=2 [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] > > stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_DISK=8:48 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_NUMBER=5 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_OFFSET=1001472 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SCHEME=dos > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SIZE=467859456 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_TYPE=0x83 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_UUID=9a4155cb-05 > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,216][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=9a4155cb > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout MINOR=53 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout PARTN=5 > > [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,217][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,218][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4429765 > > [2021-11-12 21:23:56,218][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd1 -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:56,282][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd1 > > [2021-11-12 21:23:56,288][ceph_volume.process][INFO ] stdout NAME="sdd1" > > KNAME="sdd1" MAJ:MIN="8:49" FSTYPE="ext2" MOUNTPOINT="" LABEL="" > > UUID="47c8d1ee-1c50-4af6-8fd5-001583a6f71f" RO="0" RM="0" MODEL="" > > SIZE="487M" STATE="" OWNER="root" GROUP="disk" MODE="brw-rw----" > > ALIGNMENT="0" PHY- SEC="512" LOG-SEC="512" ROTA="0" SCHED="mq-deadline" > > TYPE="part" DISC-ALN="0" DISC-GRAN="512B" DISC-MAX="2G" DISC-ZERO="0" > > PKNAME="sdd" PARTLABEL="" [2021-11-12 > > 21:23:56,289][ceph_volume.process][INFO ] Running command: /usr/ > > sbin/blkid -c /dev/null -p /dev/sdd1 > > [2021-11-12 21:23:56,293][ceph_volume.process][INFO ] stdout /dev/sdd1: > > UUID="47c8d1ee-1c50-4af6-8fd5-001583a6f71f" VERSION="1.0" > > BLOCK_SIZE="1024" TYPE="ext2" USAGE="filesystem" PART_ENTRY_SCHEME="dos" > > PART_ENTRY_UUID="9a4155cb-01" PART_ENTRY_TYPE="0x83" > > PART_ENTRY_FLAGS="0x80" PART_ENTRY_NUMBER="1" PART_ENTRY_OFFSET="2048" > > PART_ENTRY_SIZE="997376" PART_ENTRY_DISK="8:48" > > [2021-11-12 21:23:56,294][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd1 > > [2021-11-12 21:23:56,350][ceph_volume.process][INFO ] stderr Failed to > > find physical volume "/dev/sdd1". > > [2021-11-12 21:23:56,350][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd1 to check for BlueStore label > > [2021-11-12 21:23:56,350][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,351][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd1 to check for BlueStore label > > [2021-11-12 21:23:56,351][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,351][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd1 > > [2021-11-12 21:23:56,364][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-partuuid/9a4155cb-01 > > /dev/disk/by-id/ata-CT240BX500SSD1_1944E3D4E7BB- part1 > > /dev/disk/by-path/pci-0000:00:1f.2-ata-5.0-part1 /dev/disk/by-uuid/ > > 47c8d1ee-1c50-4af6-8fd5-001583a6f71f > > /dev/disk/by-path/pci-0000:00:1f.2-ata-5- part1 > > [2021-11-12 21:23:56,364][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/ sdd1 > > [2021-11-12 21:23:56,364][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd/s > > dd 1 [2021-11-12 21:23:56,364][ceph_volume.process][INFO ] stdout > > DEVTYPE=partition > > [2021-11-12 21:23:56,364][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,365][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=ext2 [2021-11-12 21:23:56,366][ceph_volume.process][INFO ] > > stdout > > ID_FS_USAGE=filesystem > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_FS_UUID=47c8d1ee-1c50-4af6-8fd5-001583a6f71f > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=47c8d1ee-1c50-4af6-8fd5-001583a6f71f > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=1.0 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_DISK=8:48 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_FLAGS=0x80 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_NUMBER=1 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_OFFSET=2048 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SCHEME=dos > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SIZE=997376 > > [2021-11-12 21:23:56,367][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_TYPE=0x83 > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_UUID=9a4155cb-01 > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=9a4155cb > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,368][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,369][ceph_volume.process][INFO ] stdout MINOR=49 > > [2021-11-12 21:23:56,369][ceph_volume.process][INFO ] stdout PARTN=1 > > [2021-11-12 21:23:56,369][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,369][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,369][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4437580 > > [2021-11-12 21:23:56,369][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,370][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd2 -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:56,426][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd2 > > [2021-11-12 21:23:56,434][ceph_volume.process][INFO ] stdout NAME="sdd2" > > KNAME="sdd2" MAJ:MIN="8:50" FSTYPE="" MOUNTPOINT="" LABEL="" UUID="" > > RO="0" RM="0" MODEL="" SIZE="1K" STATE="" OWNER="root" GROUP="disk" > > MODE="brw-rw----" ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" ROTA="0" > > SCHED="mq-deadline" TYPE="part" DISC-ALN="0" DISC-GRAN="512B" > > DISC-MAX="2G" DISC-ZERO="0" PKNAME="sdd" PARTLABEL="" > > [2021-11-12 21:23:56,435][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdd2 > > [2021-11-12 21:23:56,439][ceph_volume.process][INFO ] stdout /dev/sdd2: > > PTUUID="4808244c" PTTYPE="dos" PART_ENTRY_SCHEME="dos" > > PART_ENTRY_UUID="9a4155cb-02" PART_ENTRY_TYPE="0x5" PART_ENTRY_NUMBER="2" > > PART_ENTRY_OFFSET="1001470" PART_ENTRY_SIZE="467859458" > > PART_ENTRY_DISK="8:48" [2021-11-12 > > 21:23:56,440][ceph_volume.process][INFO > > ] Running command: /usr/ sbin/pvs --noheadings --readonly --units=b > > --nosuffix --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd2 > > [2021-11-12 21:23:56,501][ceph_volume.process][INFO ] stderr Cannot use > > /dev/ sdd2: device is too small (pv_min_size) > > [2021-11-12 21:23:56,502][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd2 to check for BlueStore label > > [2021-11-12 21:23:56,502][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,503][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd2 to check for BlueStore label > > [2021-11-12 21:23:56,503][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,503][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd2 > > [2021-11-12 21:23:56,513][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-partuuid/9a4155cb-02 > > /dev/disk/by-id/ata-CT240BX500SSD1_1944E3D4E7BB- part2 > > /dev/disk/by-path/pci-0000:00:1f.2-ata-5.0-part2 /dev/disk/by-path/ > > pci-0000:00:1f.2-ata-5-part2 > > [2021-11-12 21:23:56,513][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/ sdd2 > > [2021-11-12 21:23:56,513][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd/s > > dd 2 [2021-11-12 21:23:56,513][ceph_volume.process][INFO ] stdout > > DEVTYPE=partition > > [2021-11-12 21:23:56,513][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,513][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,514][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,515][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_DISK=8:48 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_NUMBER=2 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_OFFSET=1001470 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SCHEME=dos > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SIZE=467859458 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_TYPE=0x5 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_UUID=9a4155cb-02 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=4808244c > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,516][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout MINOR=50 > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout PARTN=2 > > [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,517][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,517][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4428059 > > [2021-11-12 21:23:56,518][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd5 -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:56,570][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd5 > > [2021-11-12 21:23:56,576][ceph_volume.process][INFO ] stdout NAME="sdd5" > > KNAME="sdd5" MAJ:MIN="8:53" FSTYPE="crypto_LUKS" MOUNTPOINT="" LABEL="" > > UUID="fbfc2e93-1c31-469b-80ce-0805c065be6f" RO="0" RM="0" MODEL="" > > SIZE="223.1G" STATE="" OWNER="root" GROUP="disk" MODE="brw-rw----" > > ALIGNMENT="0" PHY-SEC="512" LOG-SEC="512" ROTA="0" SCHED="mq-deadline" > > TYPE="part" DISC-ALN="0" DISC-GRAN="512B" DISC-MAX="2G" DISC-ZERO="0" > > PKNAME="sdd" PARTLABEL="" > > [2021-11-12 21:23:56,577][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/blkid -c /dev/null -p /dev/sdd5 > > [2021-11-12 21:23:56,580][ceph_volume.process][INFO ] stdout /dev/sdd5: > > VERSION="2" UUID="fbfc2e93-1c31-469b-80ce-0805c065be6f" > > TYPE="crypto_LUKS" > > USAGE="crypto" PART_ENTRY_SCHEME="dos" PART_ENTRY_UUID="9a4155cb-05" > > PART_ENTRY_TYPE="0x83" PART_ENTRY_NUMBER="5" PART_ENTRY_OFFSET="1001472" > > PART_ENTRY_SIZE="467859456" PART_ENTRY_DISK="8:48" > > [2021-11-12 21:23:56,581][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd5 > > [2021-11-12 21:23:56,650][ceph_volume.process][INFO ] stderr Failed to > > find physical volume "/dev/sdd5". > > [2021-11-12 21:23:56,650][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd5 to check for BlueStore label > > [2021-11-12 21:23:56,651][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,651][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd5 to check for BlueStore label > > [2021-11-12 21:23:56,651][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,651][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd5 > > [2021-11-12 21:23:56,664][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-path/pci-0000:00:1f.2-ata-5-part5 > > /dev/disk/by-partuuid/9a4155cb-05 / > > dev/disk/by-id/ata-CT240BX500SSD1_1944E3D4E7BB-part5 /dev/disk/by-uuid/ > > fbfc2e93-1c31-469b-80ce-0805c065be6f /dev/disk/by-path/pci-0000:00:1f.2- > > ata-5.0-part5 > > [2021-11-12 21:23:56,664][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/ sdd5 > > [2021-11-12 21:23:56,664][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd/s > > dd 5 [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > DEVTYPE=partition > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,665][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,666][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=crypto_LUKS > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_FS_USAGE=crypto > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_FS_UUID=fbfc2e93-1c31-469b-80ce-0805c065be6f > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=fbfc2e93-1c31-469b-80ce-0805c065be6f > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=2 [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] > > stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_DISK=8:48 > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_NUMBER=5 > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_OFFSET=1001472 > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SCHEME=dos > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SIZE=467859456 > > [2021-11-12 21:23:56,667][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_TYPE=0x83 > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_UUID=9a4155cb-05 > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=9a4155cb > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,668][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,669][ceph_volume.process][INFO ] stdout MINOR=53 > > [2021-11-12 21:23:56,669][ceph_volume.process][INFO ] stdout PARTN=5 > > [2021-11-12 21:23:56,669][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,669][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,669][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4429765 > > [2021-11-12 21:23:56,670][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/lvs --noheadings --readonly --separator=";" -a --units=b > > --nosuffix -S lv_path=/dev/sdd1 -o > > lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size [2021-11-12 > > 21:23:56,730][ceph_volume.process][INFO ] Running command: /usr/ > > bin/lsblk --nodeps -P -o > > NAME,KNAME,MAJ:MIN,FSTYPE,MOUNTPOINT,LABEL,UUID,RO,RM,MODEL,SIZE,STATE,OWN > > ER ,GROUP,MODE,ALIGNMENT,PHY- > > SEC,LOG-SEC,ROTA,SCHED,TYPE,DISC-ALN,DISC-GRAN,DISC-MAX,DISC- > > ZERO,PKNAME,PARTLABEL /dev/sdd1 > > [2021-11-12 21:23:56,738][ceph_volume.process][INFO ] stdout NAME="sdd1" > > KNAME="sdd1" MAJ:MIN="8:49" FSTYPE="ext2" MOUNTPOINT="" LABEL="" > > UUID="47c8d1ee-1c50-4af6-8fd5-001583a6f71f" RO="0" RM="0" MODEL="" > > SIZE="487M" STATE="" OWNER="root" GROUP="disk" MODE="brw-rw----" > > ALIGNMENT="0" PHY- SEC="512" LOG-SEC="512" ROTA="0" SCHED="mq-deadline" > > TYPE="part" DISC-ALN="0" DISC-GRAN="512B" DISC-MAX="2G" DISC-ZERO="0" > > PKNAME="sdd" PARTLABEL="" [2021-11-12 > > 21:23:56,738][ceph_volume.process][INFO ] Running command: /usr/ > > sbin/blkid -c /dev/null -p /dev/sdd1 > > [2021-11-12 21:23:56,744][ceph_volume.process][INFO ] stdout /dev/sdd1: > > UUID="47c8d1ee-1c50-4af6-8fd5-001583a6f71f" VERSION="1.0" > > BLOCK_SIZE="1024" TYPE="ext2" USAGE="filesystem" PART_ENTRY_SCHEME="dos" > > PART_ENTRY_UUID="9a4155cb-01" PART_ENTRY_TYPE="0x83" > > PART_ENTRY_FLAGS="0x80" PART_ENTRY_NUMBER="1" PART_ENTRY_OFFSET="2048" > > PART_ENTRY_SIZE="997376" PART_ENTRY_DISK="8:48" > > [2021-11-12 21:23:56,745][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/pvs --noheadings --readonly --units=b --nosuffix > > --separator=";" -o > > vg_name,pv_count,lv_count,vg_attr,vg_extent_count,vg_free_count,vg_extent_ > > s ize /dev/sdd1 > > [2021-11-12 21:23:56,814][ceph_volume.process][INFO ] stderr Failed to > > find physical volume "/dev/sdd1". > > [2021-11-12 21:23:56,814][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd1 to check for BlueStore label > > [2021-11-12 21:23:56,815][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,815][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd1 to check for BlueStore label > > [2021-11-12 21:23:56,815][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,815][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd1 > > [2021-11-12 21:23:56,828][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-id/ata-CT240BX500SSD1_1944E3D4E7BB-part1 > > /dev/disk/by-path/ pci-0000:00:1f.2-ata-5.0-part1 > > /dev/disk/by-path/pci-0000:00:1f.2-ata-5-part1 > > /dev/disk/by-uuid/47c8d1ee-1c50-4af6-8fd5-001583a6f71f > > /dev/disk/by-partuuid/ 9a4155cb-01 > > [2021-11-12 21:23:56,828][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/ sdd1 > > [2021-11-12 21:23:56,828][ceph_volume.process][INFO ] stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd/s > > dd 1 [2021-11-12 21:23:56,828][ceph_volume.process][INFO ] stdout > > DEVTYPE=partition > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,829][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,830][ceph_volume.process][INFO ] stdout > > ID_FS_TYPE=ext2 [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] > > stdout > > ID_FS_USAGE=filesystem > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_FS_UUID=47c8d1ee-1c50-4af6-8fd5-001583a6f71f > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_FS_UUID_ENC=47c8d1ee-1c50-4af6-8fd5-001583a6f71f > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_FS_VERSION=1.0 > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_DISK=8:48 > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_FLAGS=0x80 > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_NUMBER=1 > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_OFFSET=2048 > > [2021-11-12 21:23:56,831][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SCHEME=dos > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_SIZE=997376 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_TYPE=0x83 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PART_ENTRY_UUID=9a4155cb-01 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=9a4155cb > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,832][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,833][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,833][ceph_volume.process][INFO ] stdout MINOR=49 > > [2021-11-12 21:23:56,833][ceph_volume.process][INFO ] stdout PARTN=1 > > [2021-11-12 21:23:56,833][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,833][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,833][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4437580 > > [2021-11-12 21:23:56,833][ceph_volume.util.disk][INFO ] opening device > > /dev/ sdd to check for BlueStore label > > [2021-11-12 21:23:56,834][ceph_volume.process][INFO ] Running command: > > /usr/ sbin/udevadm info --query=property /dev/sdd > > [2021-11-12 21:23:56,846][ceph_volume.process][INFO ] stdout > > DEVLINKS=/dev/ disk/by-path/pci-0000:00:1f.2-ata-5.0 > > /dev/disk/by-path/pci-0000:00:1f.2-ata-5 > > /dev/disk/by-id/ata-CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,846][ceph_volume.process][INFO ] stdout > > DEVNAME=/dev/sdd [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] > > stdout DEVPATH=/ > > devices/pci0000:00/0000:00:1f.2/ata5/host4/target4:0:0/4:0:0:0/block/sdd > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > DEVTYPE=disk > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout ID_ATA=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_DOWNLOAD_MICROCODE=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_CURRENT_VALUE=254 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_APM_ENABLED=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_HPA_ENABLED=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_PM_ENABLED=1 > > [2021-11-12 21:23:56,847][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENABLED=0 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ENHANCED_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SECURITY_ERASE_UNIT_MIN=2 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_FEATURE_SET_SMART_ENABLED=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_ROTATION_RATE_RPM=0 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA=1 [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] > > stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN1=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_SATA_SIGNAL_RATE_GEN2=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout > > ID_ATA_WRITE_CACHE_ENABLED=1 > > [2021-11-12 21:23:56,848][ceph_volume.process][INFO ] stdout ID_BUS=ata > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_MODEL=CT240BX500SSD1 > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_MODEL_ENC=CT240BX500SSD1\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x2 > > 0\ x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20\x20 [2021-11-12 > > 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_TYPE=dos > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_PART_TABLE_UUID=9a4155cb > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_PATH=pci-0000:00:1f.2-ata-5.0 > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_PATH_ATA_COMPAT=pci-0000:00:1f.2-ata-5 > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_PATH_TAG=pci-0000_00_1f_2-ata-5_0 > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_REVISION=M6CR013 > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_SERIAL=CT240BX500SSD1_1944E3D4E7BB > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_SERIAL_SHORT=1944E3D4E7BB > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout > > ID_TYPE=disk > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout MAJOR=8 > > [2021-11-12 21:23:56,849][ceph_volume.process][INFO ] stdout MINOR=48 > > [2021-11-12 21:23:56,850][ceph_volume.process][INFO ] stdout > > SUBSYSTEM=block [2021-11-12 21:23:56,850][ceph_volume.process][INFO ] > > stdout TAGS=:systemd: [2021-11-12 21:23:56,850][ceph_volume.process][INFO > > ] stdout > > USEC_INITIALIZED=4425899 > > [2021-11-12 21:23:56,854][ceph_volume.util.system][INFO ] /dev/sda was > > not found as mounted > > [2021-11-12 21:23:56,861][ceph_volume.util.system][INFO ] /dev/sdb was > > not found as mounted > > [2021-11-12 21:23:56,865][ceph_volume.util.system][INFO ] /dev/sdc was > > not found as mounted > > > > > > > > On Friday, 12 November 2021 21:13:05 GMT Igor Fedotov wrote: > > > > > Hi Stephen, > > > > > > > > > > > > it would be nice to see failing OSD startup log... > > > > > > > > > > > > > > > Thanks, > > > > > > > > > > > > Igor > > > > > > > > > > > > On 11/12/2021 11:37 PM, Stephen J. Thompson wrote: > > > > > > > Before shutting down > > > > > > > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx