Re: [ceph-users] wired output of ceph df : Firefly 0.80.5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yeah, this is pretty weird. Can you run "df" on each of your OSDs and
see if any of them individually have a strange output?
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com


On Tue, Aug 12, 2014 at 6:52 AM, Karan Singh <karan.singh@xxxxxx> wrote:
> Forgot to mention i am observing  EB  in ceph -s output  ,  does it mean
> Exabyte  ;-)
>
>
> # ceph -s
>     cluster 009d3518-e60d-4f74-a26d-c08c1976263c
>      health HEALTH_WARN 'cache-pool' at/near target max
>      monmap e3: 3 mons at
>      mdsmap e14: 1/1/1 up {0=storage0101-ib=up:active}
>      osdmap e194215: 402 osds: 402 up, 402 in
>       pgmap v743051: 31168 pgs, 22 pools, 8 EB data, 378 kobjects
>             17508 GB used, 1284 TB / 1301 TB avail
>                31168 active+clean
>
>
> - Karan Singh -
>
>
> On 12 Aug 2014, at 16:45, Karan Singh <karan.singh@xxxxxx> wrote:
>
> Hello Developers
>
> I have encountered some wired output of ceph df command , suddenly
>
>
> When i was writing some data on cache-pool , and checked its used % , i
> found some used as 8E ( don’t know what is this ) and the used % for
> cache-pool was 0
>
>
> # ceph df
> GLOBAL:
>     SIZE      AVAIL     RAW USED     %RAW USED
>     1301T     1284T     17518G       1.31
> POOLS:
>     NAME                   ID     USED       %USED     OBJECTS
>     data                   0      801M       0         2
>     metadata               1      801M       0         22
>     rbd                    2      0          0         0
>     .rgw                   3      3524       0         26
>     .rgw.root              4      778        0         3
>     .rgw.control           5      0          0         8
>     .rgw.buckets           6      8201M      0         2298
>     .rgw.buckets.index     7      0          0         13
>     .users.swift           8      7          0         1
>     volumes                9      1106G      0.08      283387
>     images                 10     40960k     0         8
>     backups                11     0          0         0
>     .rgw.gc                12     0          0         32
>     .users.uid             13     848        0         5
>     .users                 14     16         0         2
>     .log                   15     153k       0         37
>                            16     0          0         0
>     hpsl4540               21     110G       0         28152
>     hpdl380                22     245G       0.02      62688
>     EC-2-2                 23     6338G      0.48      4859
>     cache-pool             24     8E         0         5849        ## What
> is the meaning of E here , also please note used % for cache-pool is 0 here
>     ssd                    25     25196M     0         5464
>
>
> After some time when cache-poo used value changed to 7E and used % as
> 644301.19  . While there were no objects in the cache-pool
>
>
> # ceph df
> GLOBAL:
>     SIZE      AVAIL     RAW USED     %RAW USED
>     1301T     1284T     17508G       1.31
> POOLS:
>     NAME                   ID     USED       %USED         OBJECTS
>     data                   0      801M       0             2
>     metadata               1      801M       0             22
>     rbd                    2      0          0             0
>     .rgw                   3      3524       0             26
>     .rgw.root              4      778        0             3
>     .rgw.control           5      0          0             8
>     .rgw.buckets           6      8201M      0             2298
>     .rgw.buckets.index     7      0          0             13
>     .users.swift           8      7          0             1
>     volumes                9      1106G      0.08          283387
>     images                 10     40960k     0             8
>     backups                11     0          0             0
>     .rgw.gc                12     0          0             32
>     .users.uid             13     848        0             5
>     .users                 14     16         0             2
>     .log                   15     153k       0             37
>                            16     0          0             0
>     hpsl4540               21     110G       0             28152
>     hpdl380                22     245G       0.02          62688
>     EC-2-2                 23     6338G      0.48          4843
>     cache-pool             24     7E         644301.19     1056
> ## The used % for cache-pool has become 644301.19
>     ssd                    25     25196M     0             5464
> #
>
>
> # rados -p cache-pool ls
> #
>
>
>
>
> Is this a bug , if yes , then is is already known. Do you want me to raise a
> bug ticket in tracker.ceph.com ?
>
>
>
> ****************************************************************
> Karan Singh
> Cloud computing group
> CSC - IT Center for Science,
> Keilaranta 14, P. O. Box 405, FIN-02101 Espoo, Finland
> tel. +358 9 4572001
> fax +358 9 4572302
> http://www.csc.fi/
> ****************************************************************
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux