The OSD even when it's down, I can still access it's contents, looks like I need to check out ceph-objectstore-tool.
# id weight type name up/down reweight
-1 98.44 root default
-2 32.82 host ceph-node-1
0 3.64 osd.0 up 1
1 3.64 osd.1 up 1
2 3.64 osd.2 up 1
3 3.64 osd.3 up 1
4 3.64 osd.4 up 1
5 3.64 osd.5 up 1
6 3.7 osd.6 up 1
7 3.64 osd.7 up 1
8 3.64 osd.8 up 1
-3 32.74 host ceph-node-2
10 3.64 osd.10 down 1
11 3.64 osd.11 up 1
12 3.64 osd.12 up 1
13 3.64 osd.13 up 1
15 3.64 osd.15 up 1
16 3.64 osd.16 up 1
17 3.64 osd.17 up 1
27 3.63 osd.27 up 1
14 3.63 osd.14 DNE
-4 32.88 host ceph-node-3
18 3.7 osd.18 up 1
19 3.7 osd.19 up 1
20 3.64 osd.20 up 1
21 3.64 osd.21 up 1
22 3.64 osd.22 up 1
23 3.64 osd.23 up 1
24 3.64 osd.24 up 1
25 3.64 osd.25 up 1
26 3.64 osd.26 up 1
-1 98.44 root default
-2 32.82 host ceph-node-1
0 3.64 osd.0 up 1
1 3.64 osd.1 up 1
2 3.64 osd.2 up 1
3 3.64 osd.3 up 1
4 3.64 osd.4 up 1
5 3.64 osd.5 up 1
6 3.7 osd.6 up 1
7 3.64 osd.7 up 1
8 3.64 osd.8 up 1
-3 32.74 host ceph-node-2
10 3.64 osd.10 down 1
11 3.64 osd.11 up 1
12 3.64 osd.12 up 1
13 3.64 osd.13 up 1
15 3.64 osd.15 up 1
16 3.64 osd.16 up 1
17 3.64 osd.17 up 1
27 3.63 osd.27 up 1
14 3.63 osd.14 DNE
-4 32.88 host ceph-node-3
18 3.7 osd.18 up 1
19 3.7 osd.19 up 1
20 3.64 osd.20 up 1
21 3.64 osd.21 up 1
22 3.64 osd.22 up 1
23 3.64 osd.23 up 1
24 3.64 osd.24 up 1
25 3.64 osd.25 up 1
26 3.64 osd.26 up 1
[root@ceph-node-2 ceph-10]# pwd
/var/lib/ceph/osd/ceph-10
[root@ceph-node-2 ceph-10]#
[root@ceph-node-2 ceph-10]# ll
total 52
-rw-r--r--. 1 root root 499 Aug 12 2014 activate.monmap
-rw-r--r--. 1 root root 3 Aug 12 2014 active
-rw-r--r--. 1 root root 37 Aug 12 2014 ceph_fsid
drwxr-xr-x. 91 root root 8192 Oct 19 11:56 current
-rw-r--r--. 1 root root 37 Aug 12 2014 fsid
lrwxrwxrwx. 1 root root 9 Feb 26 2017 journal -> /dev/sdj2
-rw-------. 1 root root 57 Aug 12 2014 keyring
-rw-r--r--. 1 root root 21 Aug 12 2014 magic
-rw-r--r--. 1 root root 6 Aug 12 2014 ready
-rw-r--r--. 1 root root 4 Aug 12 2014 store_version
-rw-r--r--. 1 root root 42 Aug 12 2014 superblock
-rw-r--r--. 1 root root 0 Oct 17 10:23 sysvinit
-rw-r--r--. 1 root root 3 Aug 12 2014 whoami
[root@ceph-node-2 ceph-10]
/var/lib/ceph/osd/ceph-10
[root@ceph-node-2 ceph-10]#
[root@ceph-node-2 ceph-10]# ll
total 52
-rw-r--r--. 1 root root 499 Aug 12 2014 activate.monmap
-rw-r--r--. 1 root root 3 Aug 12 2014 active
-rw-r--r--. 1 root root 37 Aug 12 2014 ceph_fsid
drwxr-xr-x. 91 root root 8192 Oct 19 11:56 current
-rw-r--r--. 1 root root 37 Aug 12 2014 fsid
lrwxrwxrwx. 1 root root 9 Feb 26 2017 journal -> /dev/sdj2
-rw-------. 1 root root 57 Aug 12 2014 keyring
-rw-r--r--. 1 root root 21 Aug 12 2014 magic
-rw-r--r--. 1 root root 6 Aug 12 2014 ready
-rw-r--r--. 1 root root 4 Aug 12 2014 store_version
-rw-r--r--. 1 root root 42 Aug 12 2014 superblock
-rw-r--r--. 1 root root 0 Oct 17 10:23 sysvinit
-rw-r--r--. 1 root root 3 Aug 12 2014 whoami
[root@ceph-node-2 ceph-10]
- Vlad
ᐧ
On Thu, Oct 17, 2019 at 11:56 AM Ashley Merrick <singapore@xxxxxxxxxxxxxx> wrote:
I think your better off doing the DD method, you can export and import a PG at a time (ceph-objectstore-tool)
But if the disk is failing a DD is probably your best method.---- On Thu, 17 Oct 2019 11:44:20 +0800 vladimir franciz blando <vladimir.blando@xxxxxxxxx> wrote ----Sorry for not being clear, when I say healthy disk, I mean those are already an OSD, so I need to transfer the data from the failed OSD to the other OSDs that are healthy.- Vladᐧ_______________________________________________On Thu, Oct 17, 2019 at 11:31 AM Konstantin Shalygin <k0ste@xxxxxxxx> wrote:
On 10/17/19 10:29 AM, vladimir franciz blando wrote:
> I have a not ideal setup on one of my cluster, 3 ceph nodes but
> using replication 1 on all pools (don't ask me why replication 1, it's
> a long story).
>
> So it has come to this situation that a disk keeps on crashing,
> possible a hardware failure and I need to recover from that.
>
> What's my best option for me to recover the data from the failed disk
> and transfer it to the other healthy disks?
>
> This cluster is using Firefly
`dd if=/dev/old_drive of=/dev/new_drive` I guess.
k
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx