Re: [object gateway] setting storage class does not move object to correct backing pool?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Gerdriaan,

I think actually moving an already-stored object requires a lifecycle
transition policy.  Assuming such a policy exists and matches the
object by prefix/tag/time, it would migrate during an (hopefully the
first) eligible lc processing window.

Matt

On Tue, Dec 10, 2019 at 7:44 AM Gerdriaan Mulder <gerdriaan@xxxxxxxx> wrote:
>
> Hi,
>
> If I change the storage class of an object via s3cmd, the object's
> storage class is reported as being changed. However, when inspecting
> where the objects are placed (via `rados -p <pool> ls`, see further on),
> the object seems to be retained in the original pool.
>
> The idea behind this test setup is to simulate two storage locations,
> one based on SSDs or similar flash storage, the other on slow HDDs. We
> want to be able to alter the storage location of objects on the fly,
> typically only from fast to slow storage. The object should then only
> reside on slow storage.
>
> The setup is as follows on Nautilus (Ubuntu 16.04, see
> <https://gist.github.com/mrngm/bba6ffdc545bfa52ebf79d6d2c002a6d> for the
> full dump):
>
> <<<<<<<<
> root@node1:~# ceph -s
>   health: HEALTH_OK
>
>   mon: 3 daemons, quorum node1,node3,node5 (age 12d)
>   mgr: node2(active, since 6d), standbys: node4
>   osd: 4 osds: 4 up (since 12d), 4 in (since 12d)
>   rgw: 1 daemon active (node1)
>
>   pools:   7 pools, 296 pgs
>   objects: 229 objects, 192 KiB
>   usage:   3.2 GiB used, 6.8 GiB / 10 GiB avail
>   pgs:     296 active+clean
>
> root@node1:~# ceph osd tree
> ID  CLASS WEIGHT  TYPE NAME               STATUS REWEIGHT PRI-AFF
>   -1       0.00970 root default
> -16       0.00970     datacenter nijmegen
>   -3       0.00388         host node2
>    0   hdd 0.00388             osd.0           up  1.00000 1.00000
>   -5       0.00388         host node3
>    1   hdd 0.00388             osd.1           up  1.00000 1.00000
>   -7       0.00098         host node4
>    2   ssd 0.00098             osd.2           up  1.00000 1.00000
>   -9       0.00098         host node5
>    3   ssd 0.00098             osd.3           up  1.00000 1.00000
>
> root@node1:~# ceph osd pool ls detail
> pool 1 'tier1-ssd' replicated size 2 min_size 1 crush_rule 1 object_hash
> rjenkins pg_num 128 pgp_num 128 [snip] application rgw
> pool 2 'tier2-hdd' replicated size 1 min_size 1 crush_rule 2 object_hash
> rjenkins pg_num 128 pgp_num 128 [snip] application rgw
> pool 3 '.rgw.root' replicated size 2 min_size 1 crush_rule 0 object_hash
> rjenkins pg_num 8 pgp_num 8 [snip] application rgw
> pool 4 'default.rgw.control' replicated size 2 min_size 1 crush_rule 0
> [snip] application rgw
> pool 5 'default.rgw.meta' replicated size 2 min_size 1 crush_rule 0
> [snip] application rgw
> pool 6 'default.rgw.log' replicated size 2 min_size 1 crush_rule 0
> [snip] application rgw
> pool 7 'default.rgw.buckets.index' replicated size 3 min_size 2
> crush_rule 0 [snip] application rgw
>
> root@node1:~# ceph osd pool application get    # compacted
>    tier1-ssd => rgw {}
>    tier2-hdd => rgw {}
>    .rgw.root => rgw {}
>     default.rgw.control => rgw {}
>     default.rgw.meta => rgw {}
>     default.rgw.log => rgw {}
>     default.rgw.buckets.index => rgw {}
>
> root@node1:~# radosgw-admin zonegroup placement list
> [
>      {
>          "key": "default-placement",
>          "val": {
>              "name": "default-placement",
>              "tags": [],
>              "storage_classes": [
>                  "SPINNING_RUST",
>                  "STANDARD"
>              ]
>          }
>      }
> ]
>
> root@node1:~# radosgw-admin zone placement list
> [
>      {
>          "key": "default-placement",
>          "val": {
>              "index_pool": "default.rgw.buckets.index",
>              "storage_classes": {
>                  "SPINNING_RUST": {
>                      "data_pool": "tier2-hdd"
>                  },
>                  "STANDARD": {
>                      "data_pool": "tier1-ssd"
>                  }
>              },
>              "data_extra_pool": "default.rgw.buckets.non-ec",
>              "index_type": 0
>          }
>      }
> ]
> ========
>
> I can also post the relevant s3cmd commands for putting objects and
> setting the storage class, but perhaps this is already enough
> information. Please let me know.
>
> <<<<<<<<
> root@node1:~# rados -p tier1-ssd ls
> ce2fc9ee-edc8-4dc7-a3fe-b1458c67168b.5805.1_darthvader.png
> ce2fc9ee-edc8-4dc7-a3fe-b1458c67168b.5805.1_2019-10-15-090436_1254x522_scrubbed.png
> ce2fc9ee-edc8-4dc7-a3fe-b1458c67168b.5805.1_kanariepiet.jpg
>
> root@node1:~# rados -p tier2-hdd ls
> ce2fc9ee-edc8-4dc7-a3fe-b1458c67168b.5805.1__shadow_.FEruUOZaVJXJcOG-e2tO1xcInNzoEvN_0
>
> $ s3cmd info s3://bucket/kanariepiet.jpg
> [snip]
>     Last mod:  Tue, 10 Dec 2019 08:09:58 GMT
>     Storage:   STANDARD
> [snip]
>
> $ s3cmd info s3://bucket/darthvader.png
> [snip]
>     Last mod:  Wed, 04 Dec 2019 10:35:14 GMT
>     Storage:   SPINNING_RUST
> [snip]
>
> $ s3cmd info s3://bucket/2019-10-15-090436_1254x522_scrubbed.png
> [snip]
>     Last mod:  Tue, 10 Dec 2019 10:33:24 GMT
>     Storage:   STANDARD
> [snip]
> ==========
>
> Any thoughts on what might occur here?
>
> Best regards,
> Gerdriaan Mulder
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>


-- 

Matt Benjamin
Red Hat, Inc.
315 West Huron Street, Suite 140A
Ann Arbor, Michigan 48103

http://www.redhat.com/en/technologies/storage

tel.  734-821-5101
fax.  734-769-8938
cel.  734-216-5309
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux