Well, we figured it out :)
This mailing list post fixed our problem
http://www.spinics.net/lists/ceph-users/msg24220.html
We had to mark the osds that were falsely reported as up, as down, and then restart all osd's
This mailing list post fixed our problem
http://www.spinics.net/lists/ceph-users/msg24220.html
We had to mark the osds that were falsely reported as up, as down, and then restart all osd's
Thanks!
On Tue, Jan 5, 2016 at 6:43 PM, Mike Carlson <mike@xxxxxxxxxxxx> wrote:
Hey ceph-usersWe upgraded from hammer to infernalis, stopped all osd's to change the user permissions from root to ceph, and all of our osd's are down (some say they are up, but the status says it is booting)ceph -scluster cabd1728-2eca-4e18-a581-b4885364e5a4health HEALTH_WARN4 pgs backfilling2905 pgs degraded844 pgs peering1137 pgs stale2905 pgs stuck degraded2881 pgs stuck inactive1137 pgs stuck stale4192 pgs stuck unclean2905 pgs stuck undersized2905 pgs undersized1 requests are blocked > 32 secrecovery 23553081/71720803 objects degraded (32.840%)recovery 5450050/71720803 objects misplaced (7.599%)mds cluster is degradednodown flag(s) setmonmap e1: 4 mons at {lts-mon=10.5.68.236:6789/0,lts-osd1=10.5.68.229:6789/0,lts-osd2=10.5.68.230:6789/0,lts-osd3=10.5.68.203:6789/0}election epoch 1162, quorum 0,1,2,3 lts-osd3,lts-osd1,lts-osd2,lts-monmdsmap e7102: 1/1/1 up {0=lts-osd1=up:replay}osdmap e6858: 102 osds: 30 up, 30 in; 2473 remapped pgsflags nodownpgmap v6218348: 4192 pgs, 7 pools, 31604 GB data, 23331 kobjects32968 GB used, 78757 GB / 109 TB avail23553081/71720803 objects degraded (32.840%)5450050/71720803 objects misplaced (7.599%)1430 undersized+degraded+peered442 remapped+peering439 active+undersized+degraded+remapped322 stale+active+remapped227 stale+active+undersized+degraded207 stale+undersized+degraded+peered183 activating+undersized+degraded165 peering159 active+undersized+degraded123 stale+peering119 activating+undersized+degraded+remapped114 stale+remapped+peering107 active+remapped59 stale+activating+undersized+degraded57 stale+active+undersized+degraded+remapped21 stale+activating+undersized+degraded+remapped6 activating+remapped6 stale+activating+remapped4 undersized+degraded+remapped+backfilling+peered1 stale+remapped1 remappedceph osd treeID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY-1 371.27994 root default-2 123.75998 host lts-osd10 3.64000 osd.0 down 0 1.000001 3.64000 osd.1 down 0 1.000002 3.64000 osd.2 down 0 1.000003 3.64000 osd.3 down 0 1.000004 3.64000 osd.4 down 0 1.000005 3.64000 osd.5 down 0 1.000006 3.64000 osd.6 down 0 1.000007 3.64000 osd.7 down 0 1.000008 3.64000 osd.8 down 0 1.000009 3.64000 osd.9 down 0 1.0000010 3.64000 osd.10 down 0 1.0000011 3.64000 osd.11 down 0 1.0000012 3.64000 osd.12 down 0 1.0000013 3.64000 osd.13 down 0 1.0000014 3.64000 osd.14 down 0 1.0000015 3.64000 osd.15 down 0 1.0000016 3.64000 osd.16 down 0 1.0000017 3.64000 osd.17 down 0 1.0000018 3.64000 osd.18 down 0 1.0000019 3.64000 osd.19 down 0 1.0000020 3.64000 osd.20 down 0 1.0000021 3.64000 osd.21 down 0 1.0000022 3.64000 osd.22 down 0 1.0000023 3.64000 osd.23 down 0 1.0000024 3.64000 osd.24 down 0 1.0000025 3.64000 osd.25 down 0 1.0000026 3.64000 osd.26 down 0 1.0000027 3.64000 osd.27 down 0 1.0000028 3.64000 osd.28 down 0 1.0000029 3.64000 osd.29 down 0 1.0000030 3.64000 osd.30 down 0 1.0000031 3.64000 osd.31 down 0 1.0000032 3.64000 osd.32 down 0 1.0000033 3.64000 osd.33 down 0 1.00000-3 123.75998 host lts-osd234 3.64000 osd.34 down 0 1.0000035 3.64000 osd.35 down 0 1.0000036 3.64000 osd.36 down 0 1.0000037 3.64000 osd.37 down 0 1.0000038 3.64000 osd.38 down 0 1.0000039 3.64000 osd.39 down 0 1.0000040 3.64000 osd.40 down 0 1.0000041 3.64000 osd.41 down 0 1.0000042 3.64000 osd.42 down 0 1.0000043 3.64000 osd.43 down 0 1.0000044 3.64000 osd.44 down 0 1.0000045 3.64000 osd.45 down 0 1.0000046 3.64000 osd.46 down 0 1.0000047 3.64000 osd.47 down 0 1.0000048 3.64000 osd.48 down 0 1.0000049 3.64000 osd.49 down 0 1.0000050 3.64000 osd.50 down 0 1.0000051 3.64000 osd.51 down 0 1.0000052 3.64000 osd.52 down 0 1.0000053 3.64000 osd.53 down 0 1.0000054 3.64000 osd.54 down 0 1.0000055 3.64000 osd.55 down 0 1.0000056 3.64000 osd.56 down 0 1.0000057 3.64000 osd.57 down 0 1.0000058 3.64000 osd.58 down 0 1.0000059 3.64000 osd.59 down 0 1.0000060 3.64000 osd.60 down 0 1.0000061 3.64000 osd.61 down 0 1.0000062 3.64000 osd.62 down 0 1.0000063 3.64000 osd.63 down 0 1.0000064 3.64000 osd.64 down 0 1.0000065 3.64000 osd.65 down 0 1.0000066 3.64000 osd.66 down 0 1.0000067 3.64000 osd.67 down 0 1.00000-4 123.75998 host lts-osd368 3.64000 osd.68 down 0 1.0000069 3.64000 osd.69 down 0 1.0000070 3.64000 osd.70 down 0 1.0000071 3.64000 osd.71 down 0 1.0000072 3.64000 osd.72 up 1.00000 1.0000073 3.64000 osd.73 up 1.00000 1.0000074 3.64000 osd.74 up 1.00000 1.0000075 3.64000 osd.75 up 1.00000 1.0000076 3.64000 osd.76 up 1.00000 1.0000077 3.64000 osd.77 up 1.00000 1.0000078 3.64000 osd.78 up 1.00000 1.0000079 3.64000 osd.79 up 1.00000 1.0000080 3.64000 osd.80 up 1.00000 1.0000081 3.64000 osd.81 up 1.00000 1.0000082 3.64000 osd.82 up 1.00000 1.0000083 3.64000 osd.83 up 1.00000 1.0000084 3.64000 osd.84 up 1.00000 1.0000085 3.64000 osd.85 up 1.00000 1.0000086 3.64000 osd.86 up 1.00000 1.0000087 3.64000 osd.87 up 1.00000 1.0000088 3.64000 osd.88 up 1.00000 1.0000089 3.64000 osd.89 up 1.00000 1.0000090 3.64000 osd.90 up 1.00000 1.0000091 3.64000 osd.91 up 1.00000 1.0000092 3.64000 osd.92 up 1.00000 1.0000093 3.64000 osd.93 up 1.00000 1.0000094 3.64000 osd.94 up 1.00000 1.0000095 3.64000 osd.95 up 1.00000 1.0000096 3.64000 osd.96 up 1.00000 1.0000097 3.64000 osd.97 up 1.00000 1.0000098 3.64000 osd.98 up 1.00000 1.0000099 3.64000 osd.99 up 1.00000 1.00000100 3.64000 osd.100 up 1.00000 1.00000101 3.64000 osd.101 up 1.00000 1.00000We have rebooted the cluster, all nodes are confirmed to have the infernalis release, but nothing we do will get a osd back up and in the cluster.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com