hi yehuda, Thanks again. i will try it . Jeff. å 2010-11-03äç 11:23 +0800ïYehuda Sadeh Weinraubåéï > Originally, the client <-> mds messages contained the full path of the > files written, similar to cifs. This was a bit racy and problematic, > but generally with a little modifications to the client code and to > the protocol you'd be able to build and send that full path through > with the osd I/O operations (again, I'm not sure what can of worms > would be opened with the full-path, or whether it's actually feasible > nowadays). > The next thing you'd need is to modify the osd operations handling, > and mirror each write operation to your ext3 partition. I think that'd > be the easiest path to go, though somewhat hacky. > > Yehuda > > On Tue, Nov 2, 2010 at 7:53 PM, Jeff Wu <cpwu@xxxxxxxxxxxxx> wrote: > > Hi Yehuda, > > > > Thank you for your reply so quickly,your idea is very useful for us. > > our design is that it only use ceph as our file system, > > don't use NFS or CIFS . i will continue to study ceph codes ,thanks. > > > > Jeff.wu > > > > > > > > > > å 2010-11-03äç 10:39 +0800ïYehuda Sadeh Weinraubåéï > >> Hi, > >> > >> there are probably a few directions that you can go, but I'm not > >> sure which would be the easiest or whether it'd actually make sense. > >> Just to have an idea about what you're really looking for, why > >> wouldn't doing nfs mount over secret/ work? Is there a real need for > >> that metadata to go through the mds? > >> > >> Yehuda > >> > >> On Tue, Nov 2, 2010 at 7:18 PM, Jeff Wu <cpwu@xxxxxxxxxxxxx> wrote: > >> > hi, > >> > > >> > Any ideas or suggestions about this? or , To solve this issue,i need > >> > deeply look into what parts ceph codes ,OSD ,CRUSH ,.OSD cluster > >> > expansion .. etc ? Thanks. > >> > > >> > > >> > Jeff.Wu > >> > > >> > > >> > ---------- Forwarded message ---------- > >> > From: Jeff Wu <cpwu@xxxxxxxxxxxxx> > >> > To: "sage@xxxxxxxxxxxx" <sage@xxxxxxxxxxxx>, "ceph-devel@xxxxxxxxxxxxxxx" <ceph-devel@xxxxxxxxxxxxxxx> > >> > Date: Tue, 2 Nov 2010 14:49:58 +0800 > >> > Subject: Questions about ceph OSD > >> > Hi , > >> > > >> > I have recently been working the preliminary research whether ceph > >> > can use at our cloud computing storage system in the future , > >> > ceph is a very very excellent file system.But ,Now ,i hit a very > >> > important problem for us: > >> > > >> > Normally, as ceph described , all of the data are saved at OSDs with > >> > CRUSHMAP at ceph system. > >> > > >> > but my goal is to : > >> > 1. add a ext3 disk to ceph system ,but doesn't set it as a OSD; > >> > 2. save the normal data at the OSDs of a ceph system; > >> > 3. save the special data to a local ext3 disk at a ceph system, > >> > the ext3 disk is not set as a ceph OSD,but its infos are added to > >> > MDS and MON,when Write the special data at a ceph client ,these > >> > special data are not saved to ceph server OSDs ,but saved at the local > >> > ext3 disk. > >> > > >> > like this: > >> > At ceph client: > >> > $ mount.ceph 1.2.3.4:6789:/ /ceph > >> > $ cd /ceph > >> > $ ls > >> > cloud user secret > >> > > >> > when do writing,"cloud and user" folder data are saved to ceph OSDs , > >> > but "secret" folder data are saved to a local ext3 disk . > >> > > >> > My question is: > >> > Q1- If i add a new OSD to the ceph system ,but i do a special CRUSH > >> > placement map . Could this ways solved this problem ?,If can ,How ? > >> > > >> > Q2 - If to implement this function as a patch myself, Could you give me > >> > any ideas ? or what parts of ceph codes(server/client) that i need look > >> > into ? > >> > > >> > Thanks for all help and reply . > >> > > >> > > >> > Jeff.Wu > >> > Transoft.inc ,Sr.software engineer. > >> > > >> > > >> > > >> > > >> > > >> > -- > >> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > >> > the body of a message to majordomo@xxxxxxxxxxxxxxx > >> > More majordomo info at http://vger.kernel.org/majordomo-info.html > >> > > >> > > > > > > > -- > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > > the body of a message to majordomo@xxxxxxxxxxxxxxx > > More majordomo info at http://vger.kernel.org/majordomo-info.html > > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html