Re: running Qemu / Hypervisor AND Ceph on the same nodes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 03/30/2015 01:29 PM, Mark Nelson wrote:
> This is definitely something that we've discussed, though I don't think
> anyone has really planned out what a complete solution would look like
> including processor affinity, etc.  Before I joined inktank I worked at
> a supercomputing institute and one of the projects we worked on was to
> develop grid computing tools for bioinformatics research.  Moving
> analytics rather than the data was a big topic for us too since genomics
> data at least tends to be pretty big.
Interestingly I work for supercomputing/education research company and
we are thinking for similar use case and purpose. So are interesting to
know other people managing resources this way.

- Gurvinder
  Potentially ceph could be a very
> interesting solution for that kind of thing.
> 
> Mark
> 
> On 03/30/2015 06:20 AM, Gurvinder Singh wrote:
>> One interesting use case of combining Ceph with computing is running big
>> data jobs on ceph itself. As with CephFS coming along, you can run
>> Haddop/Spark jobs directly on ceph without needed to move your data to
>> compute resources with data locality support. I am wondering if anyone
>> in community is looking at combining storage and compute resources from
>> this point of view.
>>
>> Regards,
>> Gurvinder
>> On 03/29/2015 09:19 PM, Nick Fisk wrote:
>>> There's probably a middle ground where you get the best of both worlds.
>>> Maybe 2-4 OSD's per compute node alongside dedicated Ceph nodes. That
>>> way
>>> you get a bit of extra storage and can still use lower end CPU's, but
>>> don't
>>> have to worry so much about resource contention.
>>>
>>>> -----Original Message-----
>>>> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On
>>>> Behalf Of
>>>> Martin Millnert
>>>> Sent: 29 March 2015 19:58
>>>> To: Mark Nelson
>>>> Cc: ceph-users@xxxxxxxxxxxxxx
>>>> Subject: Re:  running Qemu / Hypervisor AND Ceph on the
>>>> same
>>>> nodes
>>>>
>>>> On Thu, Mar 26, 2015 at 12:36:53PM -0500, Mark Nelson wrote:
>>>>> Having said that, small nodes are
>>>>> absolutely more expensive per OSD as far as raw hardware and
>>>>> power/cooling goes.
>>>>
>>>> The smaller volume manufacturers have on the units, the worse the
>>>> margin
>>>> typically (from buyers side).  Also, CPUs typically run up a premium
>>>> the
>>> higher
>>>> you go.  I've found a lot of local maximas, optimization-wise, over the
>>> past
>>>> years both in 12 OSD/U vs 18 OSD/U dedicated storage node setups, for
>>>> instance.
>>>>    There may be local maximas along colocated low-scale storage/compute
>>>> nodes, but the one major problem with colocating storage with
>>>> compute is
>>>> that you can't scale compute independently from storage efficiently, on
>>>> using that building block alone.  There may be temporal
>>>> optimizations in
>>>> doing so however (e.g. before you have reached sufficient scale).
>>>>
>>>> There's no single optimal answer when you're dealing with 20+
>>>> variables to
>>>> consider... :)
>>>>
>>>> BR,
>>>> Martin
>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux