Thank you for the respond.
well, i have two VMs (client - server) and i went through the whole 5 min, the RBD t and CephFS, quick start commands. at the configuration file i remain at the default value but i put a default value for pg and pgp = 200, you can see it at the attached file.
i made the steps shown in the link:
i made the steps shown in the link:
and when trying to identify the object location i didnt get the result as expected instead i got the following message:
Pool obj_name does not exist.
Am i missing something here?
Best regards,
Best regards,
Thank you
On Tue, Mar 26, 2013 at 11:57 AM, Sebastien Han <sebastien.han@xxxxxxxxxxxx> wrote:
Hi,Could you post all the steps you made here? It would be easier to help you.No need to shout.Cheers.––––
Sébastien Han
Cloud Engineer
"Always give 100%. Unless you're giving blood."
PHONE : +33 (0)1 49 70 99 72 – MOBILE : +33 (0)6 52 84 44 70
EMAIL : sebastien.han@xxxxxxxxxxxx – SKYPE : han.sbastien
ADDRESS : 10, rue de la Victoire – 75009 Paris
WEB : www.enovance.com – TWITTER : @enovance_______________________________________________Hi,
I tried to locate an object to test the Ceph, and after trying to do the map the following msg appear : Pool obj_name does not exist
what went wrong !! and exactly where the files as objects get stored in server (path to the objects) !!!
Thank you.
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[global] # For version 0.55 and beyond, you must explicitly enable # or disable authentication with "auth" entries in [global]. auth cluster required = none auth service required = none auth client required = none # Ensure you have a realistic number of placement groups. We recommend # approximately 100 per OSD. E.g., total number of OSDs multiplied by 100 # divided by the number of replicas (i.e., osd pool default size). So for # 2 OSDs and osd pool default size = 1, we'd recommend approximately # (100 * 2) / 1 = 200. osd pool default size = 1 # Write an object 3 times. osd pool default min size = 1 # Allow writing one copy in a degraded state. osd pool default pg num = 200 osd pool default pgp num = 200 [osd] osd journal size = 1000 #The following assumes ext4 filesystem. filestore xattr use omap = true # For Bobtail (v 0.56) and subsequent versions, you may # add settings for mkcephfs so that it will create and mount # the file system on a particular OSD for you. Remove the comment `#` # character for the following settings and replace the values # in braces with appropriate values, or leave the following settings # commented out to accept the default values. You must specify the # --mkfs option with mkcephfs in order for the deployment script to # utilize the following settings, and you must define the 'devs' # option for each osd instance; see below. #osd mkfs type = {fs-type} #osd mkfs options {fs-type} = {mkfs options} # default for xfs is "-f" #osd mount options {fs-type} = {mount options} # default mount option is "rw,noatime" # For example, for ext4, the mount option might look like this: #osd mkfs options ext4 = user_xattr,rw,noatime # Execute $ hostname to retrieve the name of your host, # and replace {hostname} with the name of your host. # For the monitor, replace {ip-address} with the IP # address of your host. [mon.a] host = justcbuser-virtual-machine mon addr = 10.242.20.249:6789 [osd.0] host = justcbuser-virtual-machine # For Bobtail (v 0.56) and subsequent versions, you may # add settings for mkcephfs so that it will create and mount # the file system on a particular OSD for you. Remove the comment `#` # character for the following setting for each OSD and specify # a path to the device if you use mkcephfs with the --mkfs option. #devs = {path-to-device} [osd.1] host = justcbuser-virtual-machine #devs = {path-to-device} [mds.a] host = justcbuser-virtual-machine
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com