ceph mount not working anymore

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



[root at chefwks01 ~]# ceph --cluster us-west01 osd crush dump

{ "devices": [

        { "id": 0,

          "name": "osd.0"},

        { "id": 1,

          "name": "osd.1"},

        { "id": 2,

          "name": "osd.2"},

        { "id": 3,

          "name": "osd.3"},

        { "id": 4,

          "name": "osd.4"}],

  "types": [

        { "type_id": 0,

          "name": "osd"},

        { "type_id": 1,

          "name": "host"},

        { "type_id": 2,

          "name": "chassis"},

        { "type_id": 3,

          "name": "rack"},

        { "type_id": 4,

          "name": "row"},

        { "type_id": 5,

          "name": "pdu"},

        { "type_id": 6,

          "name": "pod"},

        { "type_id": 7,

          "name": "room"},

        { "type_id": 8,

          "name": "datacenter"},

        { "type_id": 9,

          "name": "region"},

        { "type_id": 10,

          "name": "root"}],

  "buckets": [

        { "id": -1,

          "name": "default",

          "type_id": 10,

          "type_name": "root",

          "weight": 147455,

          "alg": "straw",

          "hash": "rjenkins1",

          "items": [

                { "id": -2,

                  "weight": 29491,

                  "pos": 0},

                { "id": -3,

                  "weight": 29491,

                  "pos": 1},

                { "id": -4,

                  "weight": 29491,

                  "pos": 2},

                { "id": -5,

                  "weight": 29491,

                  "pos": 3},

                { "id": -6,

                  "weight": 29491,

                  "pos": 4}]},

        { "id": -2,

          "name": "ceph-node20",

          "type_id": 1,

          "type_name": "host",

          "weight": 29491,

          "alg": "straw",

          "hash": "rjenkins1",

          "items": [

                { "id": 0,

                  "weight": 29491,

                  "pos": 0}]},

        { "id": -3,

          "name": "ceph-node22",

          "type_id": 1,

          "type_name": "host",

          "weight": 29491,

          "alg": "straw",

          "hash": "rjenkins1",

          "items": [

                { "id": 2,

                  "weight": 29491,

                  "pos": 0}]},

        { "id": -4,

          "name": "ceph-node24",

          "type_id": 1,

          "type_name": "host",

          "weight": 29491,

          "alg": "straw",

          "hash": "rjenkins1",

          "items": [

                { "id": 4,

                  "weight": 29491,

                  "pos": 0}]},

        { "id": -5,

          "name": "ceph-node21",

          "type_id": 1,

          "type_name": "host",

          "weight": 29491,

          "alg": "straw",

          "hash": "rjenkins1",

          "items": [

                { "id": 1,

                  "weight": 29491,

                  "pos": 0}]},

        { "id": -6,

          "name": "ceph-node23",

          "type_id": 1,

          "type_name": "host",

          "weight": 29491,

          "alg": "straw",

          "hash": "rjenkins1",

          "items": [

                { "id": 3,

                  "weight": 29491,

                  "pos": 0}]}],

  "rules": [

        { "rule_id": 0,

          "rule_name": "replicated_ruleset",

          "ruleset": 0,

          "type": 1,

          "min_size": 1,

          "max_size": 10,

          "steps": [

                { "op": "take",

                  "item": -1,

                  "item_name": "default"},

                { "op": "chooseleaf_firstn",

                  "num": 0,

                  "type": "host"},

                { "op": "emit"}]},

        { "rule_id": 1,

          "rule_name": "erasure-code",

          "ruleset": 1,

          "type": 3,

          "min_size": 3,

          "max_size": 20,

          "steps": [

                { "op": "set_chooseleaf_tries",

                  "num": 5},

                { "op": "take",

                  "item": -1,

                  "item_name": "default"},

                { "op": "chooseleaf_indep",

                  "num": 0,

                  "type": "host"},

                { "op": "emit"}]},

        { "rule_id": 2,

          "rule_name": "ecpool",

          "ruleset": 2,

          "type": 3,

          "min_size": 3,

          "max_size": 20,

          "steps": [

                { "op": "set_chooseleaf_tries",

                  "num": 5},

                { "op": "take",

                  "item": -1,

                  "item_name": "default"},

                { "op": "choose_indep",

                  "num": 0,

                  "type": "osd"},

                { "op": "emit"}]}],

  "tunables": { "choose_local_tries": 0,

      "choose_local_fallback_tries": 0,

      "choose_total_tries": 50,

      "chooseleaf_descend_once": 1,

      "profile": "bobtail",

      "optimal_tunables": 0,

      "legacy_tunables": 0,

      "require_feature_tunables": 1,

      "require_feature_tunables2": 1}}


On Thu, Jul 10, 2014 at 8:16 PM, Sage Weil <sweil at redhat.com> wrote:

> That is CEPH_FEATURE_CRUSH_V2.  Can you attach teh output of
>
>  ceph osd crush dump
>
> Thanks!
> sage
>
>
> On Thu, 10 Jul 2014, Joshua McClintock wrote:
>
> > Yes, I change some of the mount options on my osds (xfs mount options),
> but
> > I think this may be the answer from dmesg, sorta looks like a version
> > mismatch:
> >
> > libceph: loaded (mon/osd proto 15/24)
> >
> > ceph: loaded (mds proto 32)
> >
> > libceph: mon0 192.168.0.14:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon0 192.168.0.14:6789 socket error on read
> >
> > libceph: mon2 192.168.0.16:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon2 192.168.0.16:6789 socket error on read
> >
> > libceph: mon1 192.168.0.15:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon1 192.168.0.15:6789 socket error on read
> >
> > libceph: mon0 192.168.0.14:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon0 192.168.0.14:6789 socket error on read
> >
> > libceph: mon2 192.168.0.16:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon2 192.168.0.16:6789 socket error on read
> >
> > libceph: mon1 192.168.0.15:6789 feature set mismatch, my 4a042aca <
> server's
> > 104a042aca, missing 1000000000
> >
> > libceph: mon1 192.168.0.15:6789 socket error on read
> >
> >
> > I maybe I didn't update as well as I thought it did.  I did hit every
> mon,
> > but I remember I couldn't upgrade to the new 'ceph' package because it
> > conflicted with 'python-ceph', so I uninstalled it (python-ceph), and
> then
> > upgraded to .80.1-2.   Maybe there's a subcomponent I missed?
> >
> >
> > Here's rpm -qa from the client:
> >
> >
> > [root at chefwks01 ~]# rpm -qa|grep ceph
> >
> > ceph-deploy-1.5.2-0.noarch
> >
> > ceph-release-1-0.el6.noarch
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> >
> > Here's rpm -qa from the mons:
> >
> >
> > [root at ceph-mon01 ~]# rpm -qa|grep ceph
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > ceph-release-1-0.el6.noarch
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> > [root at ceph-mon01 ~]#
> >
> >
> > [root at ceph-mon02 ~]# rpm -qa|grep ceph
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > ceph-release-1-0.el6.noarch
> >
> > [root at ceph-mon02 ~]#
> >
> >
> > [root at ceph-mon03 ~]# rpm -qa|grep ceph
> >
> > libcephfs1-0.80.1-0.el6.x86_64
> >
> > ceph-0.80.1-2.el6.x86_64
> >
> > ceph-release-1-0.el6.noarch
> >
> > [root at ceph-mon03 ~]#
> >
> >
> > Joshua
> >
> >
> >
> > On Thu, Jul 10, 2014 at 6:04 PM, Sage Weil <sweil at redhat.com> wrote:
> >       Have you made any other changes after the upgrade?  (Like
> >       adjusting
> >       tunables, or creating EC pools?)
> >
> >       See if there is anything in 'dmesg' output.
> >
> >       sage
> >
> >       On Thu, 10 Jul 2014, Joshua McClintock wrote:
> >
> >       > I upgraded my cluster to .80.1-2 (CentOS).  My mount command
> >       just freezes
> >       > and outputs an error:
> >       >
> >       > mount.ceph 192.168.0.14,192.168.0.15,192.168.0.16:/ /us-west01
> >       -o
> >       > name=chefwks01,secret=`ceph-authtool -p -n client.admin
> >       > /etc/ceph/us-west01.client.admin.keyring`
> >       >
> >       > mount error 5 = Input/output error
> >       >
> >       >
> >       > Here's the output from 'ceph -s'
> >       >
> >       >
> >       >     cluster xxxxxxxxxxxxxxxxxxxxxx
> >       >
> >       >      health HEALTH_OK
> >       >
> >       >      monmap e1: 3 monsat{ceph-mon01=
> 192.168.0.14:6789/0,ceph-mon02=192.168.0.15:6789/0,ceph-mon03
> >       =1
> >       > 92.168.0.16:6789/0}, election epoch 88, quorum 0,1,2
> >       > ceph-mon01,ceph-mon02,ceph-mon03
> >       >
> >       >      mdsmap e26: 1/1/1 up {0=0=up:active}
> >       >
> >       >      osdmap e1371: 5 osds: 5 up, 5 in
> >       >
> >       >       pgmap v49431: 192 pgs, 3 pools, 135 GB data, 34733
> >       objects
> >       >
> >       >             406 GB used, 1874 GB / 2281 GB avail
> >       >
> >       >                  192 active+clean
> >       >
> >       >
> >       > I can see some packets being exchanged between the client and
> >       the mon, but
> >       > it's a pretty short exchange.
> >       >
> >       > Any ideas where to look next?
> >       >
> >       > Joshua
> >       >
> >       >
> >       >
> >
> >
> >
> >
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140710/41deadc1/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux