Re: Recovering from a Failed Disk (replication 1)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The OSD even when it's down, I can still access it's contents, looks like I need to check out ceph-objectstore-tool.  

# id    weight  type name       up/down reweight
-1      98.44   root default
-2      32.82           host ceph-node-1
0       3.64                    osd.0   up      1
1       3.64                    osd.1   up      1
2       3.64                    osd.2   up      1
3       3.64                    osd.3   up      1
4       3.64                    osd.4   up      1
5       3.64                    osd.5   up      1
6       3.7                     osd.6   up      1
7       3.64                    osd.7   up      1
8       3.64                    osd.8   up      1
-3      32.74           host ceph-node-2
10      3.64                    osd.10  down    1
11      3.64                    osd.11  up      1
12      3.64                    osd.12  up      1
13      3.64                    osd.13  up      1
15      3.64                    osd.15  up      1
16      3.64                    osd.16  up      1
17      3.64                    osd.17  up      1
27      3.63                    osd.27  up      1
14      3.63                    osd.14  DNE
-4      32.88           host ceph-node-3
18      3.7                     osd.18  up      1
19      3.7                     osd.19  up      1
20      3.64                    osd.20  up      1
21      3.64                    osd.21  up      1
22      3.64                    osd.22  up      1
23      3.64                    osd.23  up      1
24      3.64                    osd.24  up      1
25      3.64                    osd.25  up      1
26      3.64                    osd.26  up      1

[root@ceph-node-2 ceph-10]# pwd
/var/lib/ceph/osd/ceph-10
[root@ceph-node-2 ceph-10]#
[root@ceph-node-2 ceph-10]# ll
total 52
-rw-r--r--.  1 root root  499 Aug 12  2014 activate.monmap
-rw-r--r--.  1 root root    3 Aug 12  2014 active
-rw-r--r--.  1 root root   37 Aug 12  2014 ceph_fsid
drwxr-xr-x. 91 root root 8192 Oct 19 11:56 current
-rw-r--r--.  1 root root   37 Aug 12  2014 fsid
lrwxrwxrwx.  1 root root    9 Feb 26  2017 journal -> /dev/sdj2
-rw-------.  1 root root   57 Aug 12  2014 keyring
-rw-r--r--.  1 root root   21 Aug 12  2014 magic
-rw-r--r--.  1 root root    6 Aug 12  2014 ready
-rw-r--r--.  1 root root    4 Aug 12  2014 store_version
-rw-r--r--.  1 root root   42 Aug 12  2014 superblock
-rw-r--r--.  1 root root    0 Oct 17 10:23 sysvinit
-rw-r--r--.  1 root root    3 Aug 12  2014 whoami
[root@ceph-node-2 ceph-10]




- Vlad


On Thu, Oct 17, 2019 at 11:56 AM Ashley Merrick <singapore@xxxxxxxxxxxxxx> wrote:
I think your better off doing the DD method, you can export and import a PG at a time (ceph-objectstore-tool)

But if the disk is failing a DD is probably your best method.



---- On Thu, 17 Oct 2019 11:44:20 +0800 vladimir franciz blando <vladimir.blando@xxxxxxxxx> wrote ----

Sorry for not being clear, when I say healthy disk, I mean those are already an OSD, so I need to transfer the data from the failed OSD to the other OSDs that are healthy.

- Vlad



On Thu, Oct 17, 2019 at 11:31 AM Konstantin Shalygin <k0ste@xxxxxxxx> wrote:

On 10/17/19 10:29 AM, vladimir franciz blando wrote:
> I have a not ideal setup on one of my cluster,  3 ceph  nodes but
> using replication 1 on all pools (don't ask me why replication 1, it's
> a long story).
>
> So it has come to this situation that a disk keeps on crashing,
> possible a hardware failure and I need to recover from that.
>
> What's my best option for me to recover the data from the failed disk
> and transfer it to the other healthy disks?
>
> This cluster is using Firefly


`dd if=/dev/old_drive of=/dev/new_drive` I guess.



k

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux