Re: Hadoop DNS/topology details

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Feb 19, 2013 at 4:39 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> On Tue, 19 Feb 2013, Noah Watkins wrote:
>> On Feb 19, 2013, at 2:22 PM, Gregory Farnum <greg@xxxxxxxxxxx> wrote:
>> > On Tue, Feb 19, 2013 at 2:10 PM, Noah Watkins <jayhawk@xxxxxxxxxxx>
>> > wrote:
>> >
>> > That is just truly annoying. Is this described anywhere in their docs?
>>
>> Not really. It's just there in the code--I can figure out the metric if
>> you're interested. I suspect it is local node, local rack, off rack
>> ordering, with no special tie breakers.
>>
>> > I don't think it would be hard to sort, if we had some mechanism for
>> > doing so (crush map nearness, presumably?),
>>
>> Topology information from the bucket hierarchy? I think it's always some
>> sort of heuristic.
>>
>> >> 1. Expand CephFS interface to return IP and hostname
>> >
>> > Ceph doesn't store hostnames anywhere ? it really can't do this. All
>> > it has is IPs associated with OSD ID numbers. :) Adding hostnames
>> > would be a monitor and map change, which we could do, but given the
>> > issues we've had with hostnames in other contexts I'd really rather
>> > not.
>>
>> What is the fate of hostnames used in ceph.conf? Could that information
>> be leveraged, when specified by the cluster admin?
>
> Those went hte way of the Dodo.

More specifically, those hostnames are used by mkcephfs (and
ceph-deploy?) for ssh'ing into the remote nodes, and they might sit in
a lot of ceph.conf's somewhere. But it's not data aggregated by the
monitors, or even used in-memory.

> However, we do have host and rack information in the crush map, at least
> for non-customized installations.  How about something like
>
>   string ceph_get_osd_crush_location(int osd, string type);
>
> or similar.  We could call that with "host" and "rack" and get exactly
> what we need, without making any changes to the data structures.

That's a good workaround, but it does rely on those fields being set
up in the CRUSH map (and makes handling cases like SSD-primary setups
a lot more challenging).
-Greg
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux