Re: Clients failing to respond to cache pressure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Paul,

we are using Kernel 4.15.0-47.

Regards 
Felix

IT-Services
Telefon 02461 61-9243
E-Mail: f.stolte@xxxxxxxxxxxxx
-------------------------------------------------------------------------------------
-------------------------------------------------------------------------------------
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt
-------------------------------------------------------------------------------------
-------------------------------------------------------------------------------------
 

Am 08.05.19, 13:58 schrieb "Paul Emmerich" <paul.emmerich@xxxxxxxx>:

    Which kernel are you using on the clients?
    
    Paul
    -- 
    Paul Emmerich
    
    Looking for help with your Ceph cluster? Contact us at https://croit.io
    
    croit GmbH
    Freseniusstr. 31h
    81247 München
    www.croit.io
    Tel: +49 89 1896585 90
    
    On Wed, May 8, 2019 at 1:10 PM Stolte, Felix <f.stolte@xxxxxxxxxxxxx> wrote:
    >
    > Hi folks,
    >
    > we are running a luminous cluster and using the cephfs for fileservices. We use Tivoli Storage Manager to backup all data in the ceph filesystem to tape for disaster recovery. Backup runs on two dedicated servers, which mounted the cephfs via kernel mount. In order to complete the Backup in time we are using 60 Backup Threads per Server. While backup is running, ceph health often changes from “OK” to “2 clients failing to respond to cache pressure”. After investigating and doing research in the mailing list I set the following parameters:
    >
    > mds_cache_memory_limit = 34359738368 (32 GB) on MDS Server
    >
    > client_oc_size = 104857600 (100 MB, default is 200 MB) on Backup Servers
    >
    > All Servers running Ubuntu 18.04 with Kernel 4.15.0-47 and ceph 12.2.11. We have 3 MDS Servers, 1 Active, 2 Standby. Changing to multiple active MDS Servers is not an option, since we are planning to use snapshots. Cephfs holds 78,815,975 files.
    >
    > Any advice on getting rid of the Warning would be very much appreciated. On a sidenote: Although MDS Cache Memory is set to 32GB htop shows 60GB Memory Usage for the ceph-mds process
    >
    > Best regards
    > Felix
    >
    > -------------------------------------------------------------------------------------
    > -------------------------------------------------------------------------------------
    > Forschungszentrum Juelich GmbH
    > 52425 Juelich
    > Sitz der Gesellschaft: Juelich
    > Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
    > Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
    > Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
    > Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
    > Prof. Dr. Sebastian M. Schmidt
    > -------------------------------------------------------------------------------------
    > -------------------------------------------------------------------------------------
    >
    >
    > _______________________________________________
    > ceph-users mailing list
    > ceph-users@xxxxxxxxxxxxxx
    > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
    

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux