Re: null characters at the end of the file on hard reboot of VM

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'm actually using ZFS as I've been a user of ZFS for 15y now, and have been saved too many times by its ability to detect issues (especially non-disk related hardware issues) that other filesystems don't.

With regards to the stripe_length, I have verified that changing the stripe_length changes the size of the resulting files, and that returning to the default value also gave the same result.

On Sun, Apr 9, 2017 at 4:51 PM Laszlo Budai <laszlo@xxxxxxxxxxxxxxxx> wrote:
Hello Kate,

thank you for your answer. Are you using ext4 or xfs in your VMs?

Kind regards,
Laszlo


On 08.04.2017 14:11, Kate Ward wrote:
> I had a similar issue where files were padded out to the length of ceph.dir.layout.stripe_unit (see email entitled "Issue with Ceph padding files out to ceph.dir.layout.stripe_unit size" on this list). I never found a solution, but I can say that I am currently running 10.2.6 and haven't seen the issue so far reappear. 10.2.5 worked as well as best I recall.
>
> I realise I'm running the next release after you, and that the circumstances aren't the same, but the patterns of behaviour are similar enough that I wanted to raise awareness.
>
> k8
>
> On Sat, Apr 8, 2017 at 6:39 AM Laszlo Budai <laszlo@xxxxxxxxxxxxxxxx <mailto:laszlo@xxxxxxxxxxxxxxxx>> wrote:
>
>     Hello Peter,
>
>     Thank you for your answer.
>     In our setup we have the virtual machines running in KVM, and accessing the ceph storage using librbd.
>     The rbd cache is set to "writethrough until flush = true". Here it is the result of ceph config show | grep cache :
>     # ceph --admin-daemon /run/ceph/guests/ceph-client.cinder.30601.140275026217152.asok config show | grep cache
>          "debug_objectcacher": "0\/5",
>          "mon_osd_cache_size": "10",
>          "mon_cache_target_full_warn_ratio": "0.66",
>          "mon_warn_on_cache_pools_without_hit_sets": "true",
>          "client_cache_size": "16384",
>          "client_cache_mid": "0.75",
>          "mds_cache_size": "100000",
>          "mds_cache_mid": "0.7",
>          "mds_dump_cache_on_map": "false",
>          "mds_dump_cache_after_rejoin": "false",
>          "osd_pool_default_cache_target_dirty_ratio": "0.4",
>          "osd_pool_default_cache_target_full_ratio": "0.8",
>          "osd_pool_default_cache_min_flush_age": "0",
>          "osd_pool_default_cache_min_evict_age": "0",
>          "osd_tier_default_cache_mode": "writeback",
>          "osd_tier_default_cache_hit_set_count": "4",
>          "osd_tier_default_cache_hit_set_period": "1200",
>          "osd_tier_default_cache_hit_set_type": "bloom",
>          "osd_tier_default_cache_min_read_recency_for_promote": "1",
>          "osd_map_cache_size": "500",
>          "osd_pg_object_context_cache_count": "64",
>          "leveldb_cache_size": "134217728",
>          "rocksdb_cache_size": "0",
>          "filestore_omap_header_cache_size": "1024",
>          "filestore_fd_cache_size": "128",
>          "filestore_fd_cache_shards": "16",
>          "keyvaluestore_header_cache_size": "4096",
>          "rbd_cache": "true",
>          "rbd_cache_writethrough_until_flush": "true",
>          "rbd_cache_size": "134217728",
>          "rbd_cache_max_dirty": "100663296",
>          "rbd_cache_target_dirty": "67108864",
>          "rbd_cache_max_dirty_age": "1",
>          "rbd_cache_max_dirty_object": "0",
>          "rbd_cache_block_writes_upfront": "false",
>          "rgw_cache_enabled": "true",
>          "rgw_cache_lru_size": "10000",
>          "rgw_keystone_token_cache_size": "10000",
>          "rgw_bucket_quota_cache_size": "10000",
>
>
>     I did some tests and the problem has appeared when I was using ext4 in the VM, but not in the case of xfs.
>     I did an other test when I was calling a sync at the end of the while loop, and in this case the issue did NOT appeared.
>
>
>     Kind regards,
>     Laszlo
>
>     On 08.04.2017 00:07, Peter Maloney wrote:
>     > You should describe your configuration...
>     >
>     > krbd? librbd? cephfs?
>     > is rbd_cache = true?
>     > rbd cache writethrough until flush = true?
>     > is it kvm?
>     > maybe the filesystem in the VM is relevant
>     >
>     > (I saw something similar testing cephfs... if I blacklisted a client and
>     > then force unmounted, I would get whole files (appended I think) or ends
>     > of files (new files) with zeros)
>     >
>     > On 04/07/17 13:36, Laszlo Budai wrote:
>     >> Hello,
>     >>
>     >> we have observed that there are null characters written into the open
>     >> files when hard rebooting a VM. Is tis a known issue?
>     >> Our VM is using ceph (0.94.10) storage.
>     >> we have a script like this:
>     >> while sleep 1; do date >> somefile ; done
>     >>
>     >> if we hard reset the VM while the above line is running we end up with
>     >> NULL characters in the file :
>     >>
>     >> 00000000  54 68 75 20 41 70 72 20  20 36 20 31 34 3a 33 37  |Thu Apr
>     >> 6 14:37|
>     >> 00000010  3a 33 33 20 43 45 53 54  20 32 30 31 37 0a 54 68  |:33 CEST
>     >> 2017.Th|
>     >> 00000020  75 20 41 70 72 20 20 36  20 31 34 3a 33 37 3a 33  |u Apr  6
>     >> 14:37:3|
>     >> 00000030  34 20 43 45 53 54 20 32  30 31 37 0a 54 68 75 20  |4 CEST
>     >> 2017.Thu |
>     >> 00000040  41 70 72 20 20 36 20 31  34 3a 33 37 3a 33 35 20  |Apr  6
>     >> 14:37:35 |
>     >> 00000050  43 45 53 54 20 32 30 31  37 0a 54 68 75 20 41 70  |CEST
>     >> 2017.Thu Ap|
>     >> 00000060  72 20 20 36 20 31 34 3a  33 37 3a 33 36 20 43 45  |r  6
>     >> 14:37:36 CE|
>     >> 00000070  53 54 20 32 30 31 37 0a  54 68 75 20 41 70 72 20  |ST
>     >> 2017.Thu Apr |
>     >> 00000080  20 36 20 31 34 3a 33 37  3a 33 39 20 43 45 53 54  | 6
>     >> 14:37:39 CEST|
>     >> 00000090  20 32 30 31 37 0a 54 68  75 20 41 70 72 20 20 36  | 2017.Thu
>     >> Apr  6|
>     >> 000000a0  20 31 34 3a 33 37 3a 34  30 20 43 45 53 54 20 32  | 14:37:40
>     >> CEST 2|
>     >> 000000b0  30 31 37 0a 54 68 75 20  41 70 72 20 20 36 20 31  |017.Thu
>     >> Apr  6 1|
>     >> 000000c0  34 3a 33 37 3a 34 31 20  43 45 53 54 20 32 30 31  |4:37:41
>     >> CEST 201|
>     >> 000000d0  37 0a 00 00 00 00 00 00  00 00 00 00 00 00 00 00
>     >> |7...............|
>     >> 000000e0  00 00 00 00 00 00 00 00  00 00 00 00 00 00 00 00
>     >> |................|
>     >>
>     >> We've observed the same in the syslog file also.
>     >>
>     >> Any thoughts about it?
>     >>
>     >> kind regards,
>     >> Laszlo
>     >> _______________________________________________
>     >> ceph-users mailing list
>     >> ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>     >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>     >
>     >
>     >
>     _______________________________________________
>     ceph-users mailing list
>     ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux