Could not find module rbd. CentOs 6.4

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes you can use other features like CephFS and Object Store on this kernel release that you are running.

- Karan Singh 


On 28 Jul 2014, at 07:45, Pratik Rupala <pratik.rupala at calsoftinc.com> wrote:

> Hi Karan,
> 
> I have basic setup of Ceph storage cluster in active+clean state on Linux kernel 2.6.32. As per your suggestion, RBD support starts from 2.6.34 kernel.
> So, can I use other facilities like object store and Cephfs on this setup with 2.6.32 or they are also not supported for this kernel version and is there any way to have Ceph block devices on Linux kernel 2.6.32?
> 
> Regards,
> Pratik Rupala
> 
> 
> On 7/25/2014 5:51 PM, Karan Singh wrote:
>> Hi Pratik
>> 
>> Ceph RBD support has been added in mainline Linux kernel starting 2.6.34 ,  The following errors shows that , RBD module is not present in kernel.
>> 
>> Its advisable to run latest stable kernel release if you need RBD to be working.
>> 
>>> ERROR: modinfo: could not find module rbd
>>> FATAL: Module rbd not found.
>>> rbd: modprobe rbd failed! (256)
>> 
>> 
>> 
>> - Karan -
>> 
>> On 25 Jul 2014, at 14:52, Pratik Rupala <pratik.rupala at calsoftinc.com> wrote:
>> 
>>> Hi,
>>> 
>>> I am deploying firefly version on CentOs 6.4. I am following quick installation instructions available at ceph.com.
>>> I have my customized kernel version in CentOs 6.4 which is 2.6.32.
>>> 
>>> I am able to create basic Ceph storage cluster with active+clean state. Now I am trying to create block device image on ceph client but it is giving messages as shown below:
>>> 
>>> [ceph at ceph-client1 ~]$ rbd create foo --size 1024
>>> 2014-07-25 22:31:48.519218 7f6721d43700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x6a7c50 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x6a8050).fault
>>> 2014-07-25 22:32:18.536771 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718006310 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f6718006580).fault
>>> 2014-07-25 22:33:09.598763 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f67180063e0 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f6718007e70).fault
>>> 2014-07-25 22:34:08.621655 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718007e70 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180080e0).fault
>>> 2014-07-25 22:35:19.581978 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718007e70 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180080e0).fault
>>> 2014-07-25 22:36:23.694665 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718007e70 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180080e0).fault
>>> 2014-07-25 22:37:28.868293 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718007e70 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180080e0).fault
>>> 2014-07-25 22:38:29.159830 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718007e70 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180080e0).fault
>>> 2014-07-25 22:39:28.854441 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718001db0 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f6718006990).fault
>>> 2014-07-25 22:40:14.581055 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718001ac0 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f671800c950).fault
>>> 2014-07-25 22:41:03.794903 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718004d30 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f671800c950).fault
>>> 2014-07-25 22:42:12.537442 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x6a4640 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x6a4a00).fault
>>> 2014-07-25 22:43:18.912430 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718008300 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180080e0).fault
>>> 2014-07-25 22:44:24.129258 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718008300 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f6718008f80).fault
>>> 2014-07-25 22:45:29.174719 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f671800a150 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f671800a620).fault
>>> 2014-07-25 22:46:34.032246 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718008390 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f671800a620).fault
>>> 2014-07-25 22:47:39.551973 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718008390 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f67180077e0).fault
>>> 2014-07-25 22:48:39.342226 7f6721b41700  0 -- 172.17.35.20:0/1003053 >> 172.17.35.22:6800/1875 pipe(0x7f6718001db0 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f6718003040).fault
>>> 
>>> I am not sure whether block device image has been created or not. Further I tried below command which fails:
>>> [ceph at ceph-client1 ~]$ sudo rbd map foo
>>> ERROR: modinfo: could not find module rbd
>>> FATAL: Module rbd not found.
>>> rbd: modprobe rbd failed! (256)
>>> 
>>> If I check the health of cluster it looks fine.
>>> [ceph at node1 ~]$ ceph -s
>>>    cluster 98f22f5d-783b-43c2-8ae7-b97a715c9c86
>>>     health HEALTH_OK
>>>     monmap e1: 1 mons at {node1=172.17.35.17:6789/0}, election epoch 1, quorum 0 node1
>>>     osdmap e5972: 3 osds: 3 up, 3 in
>>>      pgmap v20011: 192 pgs, 3 pools, 142 bytes data, 2 objects
>>>            190 MB used, 45856 MB / 46046 MB avail
>>>                 192 active+clean
>>> 
>>> Please let me know if I am doing anything wrong.
>>> 
>>> Regards,
>>> Pratik Rupala
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users at lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140728/14ca9c59/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux