Hello, I'm currently in the fase of testing a Ceph setup to see if it will fit our need for a 3 DC storage sollution. I install Centos 7 with Ceph version 10.2.2 I have a few things that I noticed so far: - In S3 radosgw-admin I see an error: [root@blsceph01-1 ~]# radosgw-admin user info --uid=testuser 2016-08-15 12:04:33.290367 7f7bea1f09c0 0 RGWZoneParams::create(): error creating default zone params: (17) File exists I have found some reference to this error online but this was related to some upgrade issue (http://tracker.ceph.com/issues/15597). My install is a fresh install of 10.2.2. I think someone else also mentioned he saw this error, but I can't find a sollution so far. - I choose not to name my cluster simply "ceph" because we could end up with multiple clusters in the future but I named my cluster blsceph01. During installation I ran into the issue that the cluster wouldn't start and I found a hard reference in the systemd files (/usr/lib/systemd/system/) to the clustername ceph (Environment=CLUSTER=ceph) and only after changing this to my clustername everything would work normally. - I currently have 3 OSD nodes with each of them 3 1TB SSD drives for OSD. So in total I have 9 OSD drives. Looking at the documentation this would give me a total of 512 PG's in total. The total number of pools that we are going to house on this storage is currently unknown, but I have started with the installation of S3 which gives me 12 pools to start with, so the pg_num and the pgp_num per pool should be set to 32. Is this correct, or am I missing something here? What if I create more pools over time and have more then 16 pools? Then my total number of PG's is higher then this number of 512. I allready see the message "too many PGs per OSD (609 > max 300)" and I could make this warning level higher, but where are the limits? - I currently have an warning stating the following: pool default.rgw.buckets.data has many more objects per pg than average (too few pgs?) Is it possible to spread the buckets in a pure S3 workload on multiple pools? Could I make a dedicated pool for a bucket if I expect that bucket to be very big, or to make a split between the buckets of different customers? Or maybe have different protection levels for different buckets? - I try to follow the following howto (http://cephnotes.ksperis.com/blog/2014/11/28/placement-pools-on-rados-gw) on how to put a bucket in a specific placement group so I can split data of different customers in different pools but some commands return an error: [root@blsceph01-1 ~]# radosgw-admin region get > region.conf.json failed to init zonegroup: (2) No such file or directory [root@blsceph01-1 ~]# radosgw-admin zone get > zone.conf.json unable to initialize zone: (2) No such file or directory This could have something to do with the other error radosgw-admin is giving me. -- Met vriendelijke groet / Best regards, Jan Hugo Prins Infra and Isilon storage consultant Better.be B.V. Auke Vleerstraat 140 E | 7547 AN Enschede | KvK 08097527 T +31 (0) 53 48 00 694 | M +31 (0)6 26 358 951 jprins@xxxxxxxxxxxx | www.betterbe.com This e-mail is intended exclusively for the addressee(s), and may not be passed on to, or made available for use by any person other than the addressee(s). Better.be B.V. rules out any and every liability resulting from any electronic transmission. _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com