Belay that, I misread your mail and thought you were talking about the counter used to assign IDs to new pools rather than the pool count reported from the PG map. John On Tue, Mar 18, 2014 at 7:12 PM, John Spray <john.spray@xxxxxxxxxxx> wrote: > Hi Matt, > > This is expected behaviour: pool IDs are not reused. > > Cheers, > John > > On Tue, Mar 18, 2014 at 6:53 PM, <Matt.Latter@xxxxxxxx> wrote: >> >> I am a novice ceph user creating a simple 4 OSD default cluster (initially) >> and experimenting with RADOS BENCH to understand basic HDD (OSD) >> performance. Each interation of rados bench -p data I want the cluster OSDs >> in initial state i.e. 0 objects . I assumed the easiest way was to remove >> and re-create the data pool each time. >> >> While this appears to work , when I run ceph -s it shows me the pool count >> is incrementing each time: >> >> matt@redstar9:~$ sudo ceph -s >> cluster c677f4c3-46a5-4ae1-b8aa-b070326c3b24 >> health HEALTH_WARN clock skew detected on mon.redstar10, mon.redstar11 >> monmap e1: 3 mons at >> {redstar10=192.168.5.40:6789/0,redstar11=192.168.5.41:6789/0,redstar9=192.168.5.39:6789/0}, >> election epoch 6, quorum 0,1,2 redstar10,redstar11,redstar9 >> osdmap e52: 4 osds: 4 up, 4 in >> pgmap v5240: 136 pgs, 14 pools, 768 MB data, 194 objects >> 1697 MB used, 14875 GB / 14876 GB avail >> 136 active+clean >> >> >> even though lspools still only shows me the 3 default pools (metadata, rbd, >> data) >> >> Is this a bug, AND/OR, is there a better way to zero my cluster for these >> experiments? >> >> Thanks, >> >> Matt Latter >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com