Emperor Upgrade: osds not starting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

Osds are not starting on any of the nodes after I upgraded ceph-0.67.4 to emperor 0.72.2. Tried to start osd see the following verbose output. The same error comes up on all nodes when starting osds.

[root@ceph2 ~]# service ceph -v start osd.20
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "user"
=== osd.20 ===
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "run dir"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "pid file"
--- ceph2# mkdir -p /var/run/ceph
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "log dir"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "auto start"
--- ceph2# [ -e /var/run/ceph/osd.20.pid ] || exit 1   # no pid, presumably not running
        pid=`cat /var/run/ceph/osd.20.pid`
        [ -e /proc/$pid ] && grep -q ceph-osd /proc/$pid/cmdline && grep -qwe -i.20 /proc/$pid/cmdline && exit 0 # running
        exit 1  # pid is something else
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "copy executable to"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd data"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "fs path"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "devs"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "btrfs devs"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "lock file"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "admin socket"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "max open files"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "restart on core dump"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "valgrind"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd crush update on start"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd crush location"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "osd crush initial weight"
/usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.20 "keyring"
--- ceph2# timeout 10 /usr/bin/ceph                     --name=osd.20                   --keyring=/var/lib/ceph/osd/ceph-20/keyring                     osd crush create-or-move                        --        20                       --- ceph2# df /var/lib/ceph/osd/ceph-20/. | tail -1 | awk '{ d= $2/1073741824 ; r = sprintf("%.2f", d); print r }'
0.45                    root=default                    host=ceph2
Invalid command:  --- doesn't represent a float
osd crush create-or-move <osdname (id|osd.id)> <float[0.0-]> <args> [<args>...] :  create entry or move existing entry for <name> <weight> at/to location <args>
Error EINVAL: invalid command
bash: line 1: 0.45: command not found
failed: 'timeout 10 /usr/bin/ceph                       --name=osd.20                   --keyring=/var/lib/ceph/osd/ceph-20/keyring                     osd crush create-or-move                        --        20                       --- ceph2# df /var/lib/ceph/osd/ceph-20/. | tail -1 | awk '{ d= $2/1073741824 ; r = sprintf("%.2f", d); print r }'
0.45                    root=default                    host=ceph2              


However, Osds start when I use ceph-osd -c /etc/ceph/ceph.conf -i <osdnum>  but not through service ceph or /etc/init.d/ceph. After I stared all the osds, ceph warning comes up with a message that a  "pool has too few pgs". I deleted the pool as there wasn't any important data in it. The same warning now comes up on a different pool.

[root@ceph1 ~]# ceph -s
    cluster c0459c67-e2cd-45f7-b580-dec1afc9dea5
     health HEALTH_WARN pool vmware-backups has too few pgs
     monmap e3: 3 mons at {a=192.168.6.101:6789/0,b=192.168.6.102:6789/0,c=192.168.6.103:6789/0}, election epoch 17684, quorum 0,1,2 a,b,c
     mdsmap e28128: 1/1/1 up {0=a=up:active}, 1 up:standby
     osdmap e7053: 30 osds: 30 up, 30 in
      pgmap v16242514: 6348 pgs, 12 pools, 9867 GB data, 2543 kobjects
            19775 GB used, 58826 GB / 78602 GB avail
                6343 active+clean
                   5 active+clean+scrubbing+deep
  client io 0 B/s rd, 617 kB/s wr, 81 op/s

[root@ceph1 ~]# ceph osd pool delete vmware-backups vmware-backups --yes-i-really-really-mean-it
pool 'vmware-backups' deleted
[root@ceph1 ~]# ceph -s
    cluster c0459c67-e2cd-45f7-b580-dec1afc9dea5
     health HEALTH_WARN pool centaur-backups has too few pgs
     monmap e3: 3 mons at {a=192.168.6.101:6789/0,b=192.168.6.102:6789/0,c=192.168.6.103:6789/0}, election epoch 17684, quorum 0,1,2 a,b,c
     mdsmap e28128: 1/1/1 up {0=a=up:active}, 1 up:standby
     osdmap e7054: 30 osds: 30 up, 30 in
      pgmap v16243076: 6048 pgs, 12 pools, 4437 GB data, 1181 kobjects
            19775 GB used, 58826 GB / 78602 GB avail
                6047 active+clean
                   1 active+clean+scrubbing+deep
  client io 54836 kB/s rd, 699 op/s


Regards,
Gagan
 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux