Thank you for your suggestion, Nick! I have re-weighted the OSDs and the status has changed to '256 active+clean'.
On Fri, Mar 20, 2015 at 10:33 AM, Nick Fisk <nick@xxxxxxxxxx> wrote:
I see the Problem, as your OSD's are only 8GB they have a zero weight, I think the minimum size you can get away with is 10GB in Ceph as the size is measured in TB and only has 2 decimal places.
For a work around try running :-
ceph osd crush reweight osd.X 1
for each osd, this will reweight the OSD's. Assuming this is a test cluster and you won't be adding any larger OSD's in the future this shouldn't cause any problems.
>
> admin@cp-admin:~/safedrive$ ceph osd tree
> # id weight type name up/down reweight
> -1 0 root default
> -2 0 host osd-001
> 0 0 osd.0 up 1
> 1 0 osd.1 up 1
> -3 0 host osd-002
> 2 0 osd.2 up 1
> 3 0 osd.3 up 1
> -4 0 host osd-003
> 4 0 osd.4 up 1
> 5 0 osd.5 up 1
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com