Sorry bout that
It's all set now, i thought that was replica count as it is also 4 and 5 :)
I can see the changes now
I can see the changes now
[root@controller-node ~]# ceph osd dump | grep 'replicated size'
pool 4 'images' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 1024 pgp_num 1024 last_change 19641 flags hashpspool stripe_width 0
pool 5 'volumes' replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 19640 flags hashpspool stripe_width 0
[root@controller-node ~]#
To my other question, will it remove the excess replicas?
/vlad
On Wed, Sep 7, 2016 at 8:51 AM, Jeff Bailey <bailey@xxxxxxxxxxx> wrote:
On 9/6/2016 8:41 PM, Vlad Blando wrote:
Hi,
My replication count now is this
[root@controller-node ~]# ceph osd lspools
4 images,5 volumes,
Those aren't replica counts they're pool ids.
[root@controller-node ~]#
and I made adjustment and made it to 3 for images and 2 to volumes to 3, it's been 30 mins now and the values did not change, how do I know if it was really changed.
this is the command I executed
ceph osd pool set images size 2
ceph osd pool set volumes size 3
ceph osd pool set images min_size 2
ceph osd pool set images min_size 2
Another question, since the previous replication count for images is 4 and volumes to 5, it will delete the excess replication right?
Thanks for the help
/vlad
ᐧ
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
ᐧ
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com