On Mon, Sep 30, 2013 at 11:50 PM, Eric Eastman <eric0e@xxxxxxx> wrote: > > Thank you for the reply > > >> -28 == -ENOSPC (No space left on device). I think it's is due to the > > fact that some osds are near full. >> >> >> Yan, Zheng > > > I thought that may be the case, but I would expect that ceph health would > tell me I had a full OSDs, but it is only saying they are near full: > > >>> # ceph health detail >>> HEALTH_WARN 9 near full osd(s) >>> osd.9 is near full at 85% >>> osd.29 is near full at 85% >>> osd.43 is near full at 91% >>> osd.45 is near full at 88% >>> osd.47 is near full at 88% >>> osd.55 is near full at 94% >>> osd.59 is near full at 94% >>> osd.67 is near full at 94% >>> osd.83 is near full at 94% > > Are these OSD's disks smaller than other OSD's. If they do, you need to lower these OSD's weights. Regards Yan, Zheng > As I still have lots of space: > > >>> # ceph df >>> GLOBAL: >>> SIZE AVAIL RAW USED %RAW USED >>> 249T 118T 131T 52.60 >>> >>> POOLS: >>> NAME ID USED %USED OBJECTS >>> data 0 0 0 0 >>> metadata 1 0 0 0 >>> rbd 2 8 0 1 >>> rbd-pool 3 67187G 26.30 17713336 > > > And I setup lots of Placement Groups: > > # ceph osd dump | grep 'rep size' | grep rbd-pool > pool 3 'rbd-pool' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins > pg_num 4500 pgp_num 4500 last_change 360 owner 0 > > Why did the OSDs fill up long before I ran out of space? > > Thanks, > > Eric > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com