Re: HSM

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This is a really useful summary from Malcolm.

In addition to the coordinator/copytool interface, there is the question of
where the policy engine gets its data from.  Lustre has the MDS changelog,
which Robinhood uses to replicate metadata into its MySQL database with all
the indices that it wants.  I don't know if CephFS MDS currently has a
similar interface.

John


On Sun, Nov 10, 2013 at 11:17 PM, Malcolm Haak <malcolm@xxxxxxx> wrote:
>
> Hi All,
>
> If you are talking specifically about Lustre HSM, its really an interface to add HSM functionality by leveraging existing HSM's (DMF for example)
>
> So with Lustre HSM you have a policy engine that triggers the migrations out of the filesystem. Rules are based around size, last accessed and target state (online, dual and offline).
>
> There is a 'coordinator' process involved here as well, it (from what I understand) runs on MDS nodes. It handles the interaction with the copytool. The copytool is provided by the HSM solution you are acutally using.
>
> For recalls when caps are aquired on the MDS for an exported file the resposible MSD contacts the coordinator, which in-turn uses the copytool to pull the required file out of the HSM.
>
> In the Lustre HSM, the objects that make up a file are all recalled and the file, not the objects, are handed to the HSM.
>
> For Lustre all it needs to keep track of is the current state of the file and the correct ID to reqest from the HSM. This is done inside the normal metadata storage.
>
> So there aren't really any hooks in that exports are triggered by the policy engine after a scan of the metadata, and the recalls are triggered when caps are requested on offline files. Then its just standard POSIX blocking until the file is available.
>
> Most of the state and ID stuff could be stored as XATTRS in cephfs. I'm not as sure how to do it for other things but as long as you could store some kind of extended metadata about whole objects, it could use the same interfaces as well.
>
> Hope that was acutually helpful and not just an obvious rehash...
>
> Regards
>
> Malcolm Haak
>
>
> On 09/11/13 18:33, Sage Weil wrote:
>>
>> The latest Lustre just added HSM support:
>>
>>         http://archive.hpcwire.com/hpcwire/2013-11-06/lustre_scores_business_class_upgrade_with_hsm.html
>>
>> Here is a slide deck with some high-level detail:
>>
>>         https://jira.hpdd.intel.com/secure/attachment/13185/Lustre_HSM_Design.pdf
>>
>> Is anyone familiar with the interfaces and requirements of the file system
>> itself?  I don't know much about how these systems are implemented, but I
>> would guess there are relatively lightweight requirements on the fs (ceph
>> mds in our case) to keep track of file state (online or archived
>> elsewhere).  And some hooks to trigger migrations?
>>
>> If anyone is interested in this area, I would be happy to help figure out
>> how to integrate things cleanly!
>>
>> sage
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux