Hi Sage, I have been working on benchmarking SMR Drives using libzbc. It appears that issuing ZBC commands for zone aware host is more inefficient as compared to normal copy operations using 'dd' command. I created a 256 MB file and placed it in memory (so that we do not have data fetch overheads). I copy this file repeatedly on a Host Aware SMR drive in 2 scenarios : a) dd - I use dumb dd that takes 1MB chunks of file and keeps copying the file to SMR drive for <writeSize> bytes. Note that dd does not take the zones into consideration. b) SMR_aware_copy - This copy takes file chunks 1MB in size, but issues ZBC commands to open each zone, write 256 MB data to the zone, close the zone, and then move to another zone till <writeSize> bytes have been written. performance results for 1GB, 10GB write sizes are 5x slower with "zone aware" writing, as compared to normal dd writing: writeSize (in GB) dd time (in min:sec) smr_aware_copy (in min:sec) 1 GB 0:7 0:34 10 GB 1:11 6:41 50 GB 5:51 NA 100 GB 11:42 NA (all writes were followed by sync command) I was trying to see if there is an internal cache of some sort in the Host Aware SMR drive, which probably serializes all writes up to certain extent for dd command, but the time for writes using the dd command for up to 100GB follow a linear pattern. I will try to see if we hit a bottleneck with dd for larger file sizes or unaligned writes. Followup questions: -------------------------- a) I think we should have some workload traces or patterns so that we can benchmark SMR drives and make allocator more SMR friendly. In particular - i) size of files, ii) alignment of files iii) % read / write/ delete workloads iv) degree of parallelism in writing. b) SMR Drive has a notion of parallel writes - i.e. multiple zones can be kept open and written to simultaneously. I do not think there are multiple heads involved but internally there is some form of "efficient parallel write to zone" mechanism in SMR. I am thinking about this because when we query SMR drive information, it shows that most effieicnt number of zones can be parallelly kept open = 128 . Maybe this is something that we can take advantage of? Thanks and Regards, Shehbaz -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html