Re: POOL_TARGET_SIZE_BYTES_OVERCOMMITTED

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dear Cephalopodians,

I can confirm the same problem described by Joe Ryner in 14.2.2. I'm also getting (in a small test setup):
-----------------------------------------------------
# ceph health detail
HEALTH_WARN 1 subtrees have overcommitted pool target_size_bytes; 1 subtrees have overcommitted pool target_size_ratio
POOL_TARGET_SIZE_BYTES_OVERCOMMITTED 1 subtrees have overcommitted pool target_size_bytes
    Pools ['rbd', '.rgw.root', 'default.rgw.control', 'default.rgw.meta', 'default.rgw.log', 'default.rgw.buckets.index', 'default.rgw.buckets.data'] overcommit available storage by 1.068x due to target_size_bytes    0  on pools []
POOL_TARGET_SIZE_RATIO_OVERCOMMITTED 1 subtrees have overcommitted pool target_size_ratio
    Pools ['rbd', '.rgw.root', 'default.rgw.control', 'default.rgw.meta', 'default.rgw.log', 'default.rgw.buckets.index', 'default.rgw.buckets.data'] overcommit available storage by 1.068x due to target_size_ratio 0.000 on pools []
-----------------------------------------------------

However, there's not much actual data STORED:
-----------------------------------------------------
# ceph df
RAW STORAGE:
    CLASS     SIZE        AVAIL       USED        RAW USED     %RAW USED 
    hdd       4.0 TiB     2.6 TiB     1.4 TiB      1.4 TiB         35.94 
    TOTAL     4.0 TiB     2.6 TiB     1.4 TiB      1.4 TiB         35.94 
 
POOLS:
    POOL                          ID     STORED      OBJECTS     USED        %USED     MAX AVAIL 
    rbd                            2     676 GiB     266.40k     707 GiB     23.42       771 GiB 
    .rgw.root                      9     1.2 KiB           4     768 KiB         0       771 GiB 
    default.rgw.control           10         0 B           8         0 B         0       771 GiB 
    default.rgw.meta              11     1.2 KiB           8     1.3 MiB         0       771 GiB 
    default.rgw.log               12         0 B         175         0 B         0       771 GiB 
    default.rgw.buckets.index     13         0 B           1         0 B         0       771 GiB 
    default.rgw.buckets.data      14     249 GiB      99.62k     753 GiB     24.57       771 GiB
-----------------------------------------------------
The main culprit here seems to be the default.rgw.buckets.data pool, but also the rbd pool contains thin images. 

As in the case of Joe, the autoscaler seems to look at the "USED" space, not at the "STORED" bytes:
-----------------------------------------------------
 POOL                         SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO  TARGET RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE 
 default.rgw.meta            1344k                3.0         4092G  0.0000                 1.0       8              on        
 default.rgw.buckets.index      0                 3.0         4092G  0.0000                 1.0       8              on        
 default.rgw.control            0                 3.0         4092G  0.0000                 1.0       8              on        
 default.rgw.buckets.data   788.6G                3.0         4092G  0.5782                 1.0     128              on        
 .rgw.root                  768.0k                3.0         4092G  0.0000                 1.0       8              on        
 rbd                        710.8G                3.0         4092G  0.5212                 1.0      64              on        
 default.rgw.log                0                 3.0         4092G  0.0000                 1.0       8              on
-----------------------------------------------------

This does seem like a bug to me. The warning actually fires on a cluster with 35 % raw usage, and things are mostly balanced. 
Is there already a tracker entry on this? 

Cheers,
	Oliver


On 2019-05-01 22:01, Joe Ryner wrote:
> I think I have figured out the issue.
> 
>  POOL        SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO  TARGET RATIO  PG_NUM  NEW PG_NUM  AUTOSCALE 
>  images    28523G                3.0        68779G  1.2441                  1000              warn 
> 
> My images are 28523G with a replication level 3 and have a total of 68779G in Raw Capacity.
> 
>  According to the documentation http://docs.ceph.com/docs/master/rados/operations/placement-groups/  
> 
> "*SIZE* is the amount of data stored in the pool. *TARGET SIZE*, if present, is the amount of data the administrator has specified that they expect to eventually be stored in this pool. The system uses the larger of the two values for its calculation.
> 
> *RATE* is the multiplier for the pool that determines how much raw storage capacity is consumed. For example, a 3 replica pool will have a ratio of 3.0, while a k=4,m=2 erasure coded pool will have a ratio of 1.5.
> 
> *RAW CAPACITY* is the total amount of raw storage capacity on the OSDs that are responsible for storing this pool’s (and perhaps other pools’) data. *RATIO* is the ratio of that total capacity that this pool is consuming (i.e., ratio = size * rate / raw capacity)."
> 
> So ratio = "28523G * 3.0/68779G" = 1.2441x
> 
> 
> So I'm oversubscribing by 1.2441x, thus the warning. 
> 
> 
> But ... looking at #ceph df
> 
> POOL         ID     STORED      OBJECTS     USED        %USED     MAX AVAIL 
> 
> images        3     9.3 TiB       2.82M      28 TiB     57.94       6.7 TiB
> 
> 
> I believe the 9.3TiB is the amount I have that is thinly provisioned vs a fully provisioned 28 TiB?
> 
> The raw capacity of the cluster is sitting at about 50% used.
> 
> 
> Shouldn't the ratio be the amount STORED(from ceph df) * SIZE (from  ceph osd pool autoscale-status) / Raw Capacity, since ceph uses thin provisioning in rbd?
> 
> Otherwise, this ratio will only work for people who don't thin provision which goes against what ceph is doing with rbd
> 
> http://docs.ceph.com/docs/master/rbd/
> 
> 
> 
> 
> 
> On Wed, May 1, 2019 at 11:44 AM Joe Ryner <jryner@xxxxxxxx <mailto:jryner@xxxxxxxx>> wrote:
> 
>     I have found a little more information.
>     When I turn off pg_autoscaler the warning goes away turn it back on and the warning comes back.
> 
>     I have ran the following:
>     # ceph osd pool autoscale-status
>      POOL        SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO  TARGET RATIO  PG_NUM  NEW PG_NUM  AUTOSCALE 
>      images    28523G                3.0        68779G  1.2441                  1000              warn      
>      locks     676.5M                3.0        68779G  0.0000                     8              warn      
>      rbd           0                 3.0        68779G  0.0000                     8              warn      
>      data          0                 3.0        68779G  0.0000                     8              warn      
>      metadata   3024k                3.0        68779G  0.0000                     8              warn      
> 
>     # ceph df
>     RAW STORAGE:
>         CLASS     SIZE       AVAIL       USED        RAW USED     %RAW USED 
>         hdd       51 TiB      26 TiB      24 TiB       24 TiB         48.15 
>         ssd       17 TiB     8.5 TiB     8.1 TiB      8.1 TiB         48.69 
>         TOTAL     67 TiB      35 TiB      32 TiB       32 TiB         48.28 
>      
>     POOLS:
>         POOL         ID     STORED      OBJECTS     USED        %USED     MAX AVAIL 
>         data          0         0 B           0         0 B         0       6.7 TiB 
>         metadata      1     6.3 KiB          21     3.0 MiB         0       6.7 TiB 
>         rbd           2         0 B           2         0 B         0       6.7 TiB 
>         images        3     9.3 TiB       2.82M      28 TiB     57.94       6.7 TiB 
>         locks         4     215 MiB         517     677 MiB         0       6.7 TiB 
> 
> 
>     It looks to me like it thinks the images pool no right in the autoscale-status.
> 
>     Below is a osd crush tree
>     # ceph osd crush tree
>     ID  CLASS WEIGHT   (compat) TYPE NAME             
>      -1       66.73337          root default          
>      -3       22.28214 22.28214     rack marack       
>      -8        7.27475  7.27475         host abacus   
>      19   hdd  1.81879  1.81879             osd.19    
>      20   hdd  1.81879  1.42563             osd.20    
>      21   hdd  1.81879  1.81879             osd.21    
>      50   hdd  1.81839  1.81839             osd.50    
>     -10        7.76500  6.67049         host gold     
>       7   hdd  0.86299  0.83659             osd.7     
>       9   hdd  0.86299  0.78972             osd.9     
>      10   hdd  0.86299  0.72031             osd.10    
>      14   hdd  0.86299  0.65315             osd.14    
>      15   hdd  0.86299  0.72586             osd.15    
>      22   hdd  0.86299  0.80528             osd.22    
>      23   hdd  0.86299  0.63741             osd.23    
>      24   hdd  0.86299  0.77718             osd.24    
>      25   hdd  0.86299  0.72499             osd.25    
>      -5        7.24239  7.24239         host hassium  
>       0   hdd  1.80800  1.52536             osd.0     
>       1   hdd  1.80800  1.65421             osd.1     
>      26   hdd  1.80800  1.65140             osd.26    
>      51   hdd  1.81839  1.81839             osd.51    
>      -2       21.30070 21.30070     rack marack2      
>     -12        7.76999  8.14474         host hamms    
>      27   ssd  0.86299  0.99367             osd.27    
>      28   ssd  0.86299  0.95961             osd.28    
>      29   ssd  0.86299  0.80768             osd.29    
>      30   ssd  0.86299  0.86893             osd.30    
>      31   ssd  0.86299  0.92583             osd.31    
>      32   ssd  0.86299  1.00227             osd.32    
>      33   ssd  0.86299  0.73099             osd.33    
>      34   ssd  0.86299  0.80766             osd.34    
>      35   ssd  0.86299  1.04811             osd.35    
>      -7        5.45636  5.45636         host parabola 
>       5   hdd  1.81879  1.81879             osd.5     
>      12   hdd  1.81879  1.81879             osd.12    
>      13   hdd  1.81879  1.81879             osd.13    
>      -6        2.63997  3.08183         host radium   
>       2   hdd  0.87999  1.05594             osd.2     
>       6   hdd  0.87999  1.10501             osd.6     
>      11   hdd  0.87999  0.92088             osd.11    
>      -9        5.43439  5.43439         host splinter 
>      16   hdd  1.80800  1.80800             osd.16    
>      17   hdd  1.81839  1.81839             osd.17    
>      18   hdd  1.80800  1.80800             osd.18    
>     -11       23.15053 23.15053     rack marack3      
>     -13        8.63300  8.98921         host helm     
>      36   ssd  0.86299  0.71931             osd.36    
>      37   ssd  0.86299  0.92601             osd.37    
>      38   ssd  0.86299  0.79585             osd.38    
>      39   ssd  0.86299  1.08521             osd.39    
>      40   ssd  0.86299  0.89500             osd.40    
>      41   ssd  0.86299  0.92351             osd.41    
>      42   ssd  0.86299  0.89690             osd.42    
>      43   ssd  0.86299  0.92480             osd.43    
>      44   ssd  0.86299  0.84467             osd.44    
>      45   ssd  0.86299  0.97795             osd.45    
>     -40        7.27515  7.89609         host samarium 
>      46   hdd  1.81879  1.90242             osd.46    
>      47   hdd  1.81879  1.86723             osd.47    
>      48   hdd  1.81879  1.93404             osd.48    
>      49   hdd  1.81879  2.19240             osd.49    
>      -4        7.24239  7.24239         host scandium 
>       3   hdd  1.80800  1.76680             osd.3     
>       4   hdd  1.80800  1.80800             osd.4     
>       8   hdd  1.80800  1.80800             osd.8     
>      52   hdd  1.81839  1.81839             osd.52    
> 
> 
>     Any ideas?
> 
> 
> 
> 
> 
>     On Wed, May 1, 2019 at 9:32 AM Joe Ryner <jryner@xxxxxxxx <mailto:jryner@xxxxxxxx>> wrote:
> 
>         Hi,
> 
>         I have an old ceph cluster and have upgraded recently from Luminous to Nautilus.  After converting to Nautilus I decided it was time to convert to bluestore.
> 
>         Before I converted the cluster was healthy but after I have a HEALTH_WARN
> 
>         #ceph health detail
>         HEALTH_WARN 1 subtrees have overcommitted pool target_size_bytes; 1 subtrees have overcommitted pool target_size_ratio
>         POOL_TARGET_SIZE_BYTES_OVERCOMMITTED 1 subtrees have overcommitted pool target_size_bytes
>             Pools ['data', 'metadata', 'rbd', 'images', 'locks'] overcommit available storage by 1.244x due to target_size_bytes    0  on pools []
>         POOL_TARGET_SIZE_RATIO_OVERCOMMITTED 1 subtrees have overcommitted pool target_size_ratio
>             Pools ['data', 'metadata', 'rbd', 'images', 'locks'] overcommit available storage by 1.244x due to target_size_ratio 0.000 on pools []
> 
>         I started with a target_size ratio of .85 on the images pool and reduced it to 0 to hopefully get the warning to go away.  The cluster seems to be running fine, I just can't figure out what the problem is and how to make the message go away.  I restarted the monitors this morning in hopes to fix it.  Anyone have any ideas?
> 
>         Thanks in advance
> 
> 
>         -- 
>         Joe Ryner
>         Associate Director
>         Center for the Application of Information Technologies (CAIT) - http://www.cait.org
>         Western Illinois University - http://www.wiu.edu
> 
> 
>         P: (309) 298-1804
>         F: (309) 298-2806
> 
> 
> 
>     -- 
>     Joe Ryner
>     Associate Director
>     Center for the Application of Information Technologies (CAIT) - http://www.cait.org
>     Western Illinois University - http://www.wiu.edu
> 
> 
>     P: (309) 298-1804
>     F: (309) 298-2806
> 
> 
> 
> -- 
> Joe Ryner
> Associate Director
> Center for the Application of Information Technologies (CAIT) - http://www.cait.org
> Western Illinois University - http://www.wiu.edu
> 
> 
> P: (309) 298-1804
> F: (309) 298-2806
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 


Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux