Re: OSD daemon changes port no

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 26 Nov 2012, hemant surale wrote:
> While I was using "cephfs" following error is observed -
> ------------------------------------------------------------------------------------------------
> root@hemantsec-virtual-machine:~# cephfs /mnt/ceph/a --pool 3
> invalid command

Try

 cephfs /mnt/ceph/a set_layout --pool 3

(set_layout is the command)

sage

> usage: cephfs path command [options]*
> Commands:
>    show_layout    -- view the layout information on a file or dir
>    set_layout     -- set the layout on an empty file,
>                      or the default layout on a directory
>    show_location  -- view the location information on a file
> Options:
>    Useful for setting layouts:
>    --stripe_unit, -u:  set the size of each stripe
>    --stripe_count, -c: set the number of objects to stripe across
>    --object_size, -s:  set the size of the objects to stripe across
>    --pool, -p:         set the pool to use
> 
>    Useful for getting location data:
>    --offset, -l:       the offset to retrieve location data for
> 
> ------------------------------------------------------------------------------------------------
> It may be silly question but unable to figure it out.
> 
> :(
> 
> 
> 
> 
> On Wed, Nov 21, 2012 at 8:59 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> > On Wed, 21 Nov 2012, hemant surale wrote:
> >> > Oh I see.  Generally speaking, the only way to guarantee separation is to
> >> > put them in different pools and distribute the pools across different sets
> >> > of OSDs.
> >>
> >> yeah that was correct approach but i found problem doing so from
> >> abstract level i.e. when I put file inside mounted dir
> >> "/home/hemant/cephfs " ( mounted using "mount.ceph" cmd ) . At that
> >> time anyways ceph is going to use default pool data to store files (
> >> here files were striped into different objects and then sent to
> >> appropriate osd ) .
> >>    So how to tell ceph to use different pools in this case ?
> >>
> >> Goal : separate read and write operations , where read will be done
> >> from one group of OSD and write is done to other group of OSD.
> >
> > First create the other pool,
> >
> >  ceph osd pool create <name>
> >
> > and then adjust the CRUSH rule to distributed to a different set of OSDs
> > for that pool.
> >
> > To allow cephfs use it,
> >
> >  ceph mds add_data_pool <poolid>
> >
> > and then:
> >
> >  cephfs /mnt/ceph/foo --pool <poolid>
> >
> > will set the policy on the directory such that new files beneath that
> > point will be stored in a different pool.
> >
> > Hope that helps!
> > sage
> >
> >
> >>
> >>
> >>
> >>
> >> -
> >> Hemant Surale.
> >>
> >>
> >> On Wed, Nov 21, 2012 at 12:33 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> >> > On Wed, 21 Nov 2012, hemant surale wrote:
> >> >> Its a little confusing question I believe .
> >> >>
> >> >> Actually there are two files X & Y.  When I am reading X from its
> >> >> primary .I want to make sure simultaneous writing of Y should go to
> >> >> any other OSD except primary OSD for X (from where my current read is
> >> >> getting served ) .
> >> >
> >> > Oh I see.  Generally speaking, the only way to guarantee separation is to
> >> > put them in different pools and distribute the pools across different sets
> >> > of OSDs.  Otherwise, it's all (pseudo)random and you never know.  Usually,
> >> > they will be different, particularly as the cluster size increases, but
> >> > sometimes they will be the same.
> >> >
> >> > sage
> >> >
> >> >
> >> >>
> >> >>
> >> >> -
> >> >> Hemant Sural.e
> >> >>
> >> >> On Wed, Nov 21, 2012 at 11:50 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> >> >> > On Wed, 21 Nov 2012, hemant surale wrote:
> >> >> >> >>    and one more thing how can it be possible to read from one osd and
> >> >> >> >> then simultaneous write to direct on other osd with less/no traffic?
> >> >> >> >
> >> >> >> > I'm not sure I understand the question...
> >> >> >>
> >> >> >> Scenario :
> >> >> >>        I have written file X.txt on some osd which is primary for filr
> >> >> >> X.txt ( direct write operation using rados cmd) .
> >> >> >>        Now while read on file X.txt is in progress, Can I make sure
> >> >> >> the simultaneous write request must be directed to other osd using
> >> >> >> crushmaps/other way?
> >> >> >
> >> >> > Nope.  The object location is based on the name.  Reads and writes go to
> >> >> > the same location so that a single OSD can serialize request.  That means,
> >> >> > for example, that a read that follows a write returns the just-written
> >> >> > data.
> >> >> >
> >> >> > sage
> >> >> >
> >> >> >
> >> >> >> Goal of task :
> >> >> >>        Trying to avoid read - write clashes as much as possible to
> >> >> >> achieve faster operations (I/O) . Although CRUSH selects osd for data
> >> >> >> placement based on pseudo random function.  is it possible ?
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >> -
> >> >> >> Hemant Surale.
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >> On Tue, Nov 20, 2012 at 10:15 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> >> >> >> > On Tue, 20 Nov 2012, hemant surale wrote:
> >> >> >> >> Hi Community,
> >> >> >> >>    I have question about port number used by ceph-osd daemon . I
> >> >> >> >> observed traffic (inter -osd communication while data ingest happened)
> >> >> >> >> on port 6802 and then after some time when I ingested second file
> >> >> >> >> after some delay port no 6804 was used . Is there any specific reason
> >> >> >> >> to change port no here?
> >> >> >> >
> >> >> >> > The ports are dynamic.  Daemons bind to a random (6800-6900) port on
> >> >> >> > startup and communicate on that.  They discover each other via the
> >> >> >> > addresses published in the osdmap when the daemon starts.
> >> >> >> >
> >> >> >> >>    and one more thing how can it be possible to read from one osd and
> >> >> >> >> then simultaneous write to direct on other osd with less/no traffic?
> >> >> >> >
> >> >> >> > I'm not sure I understand the question...
> >> >> >> >
> >> >> >> > sage
> >> >> >> --
> >> >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >> >>
> >> >> >>
> >> >>
> >> >>
> >>
> >>
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux