Re: iozone test crashed on ceph

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Jun 4, 2012 at 4:43 PM, udit agarwal <fzdudit@xxxxxxxxx> wrote:
> Sorry ,the link is: [...]

If you run iozone again, does the bug happen again?

Comparing your iozone run with our test suite, we don't currently do
-i 2 (random read/write), and we only run a few specific record sizes
to save time (-r 16K,1M); your run is talking about record len "4"
(K?) when it crashes. That might be relevant for triggering this.

Summarizing for others:

Jun  4 22:19:03 hp1 kernel: [ 7627.132026] BUG: unable to handle
kernel NULL pointer dereference at 000000000000000a
Jun  4 22:19:03 hp1 kernel: [ 7627.132036] IP: [<ffffffffa01a1fd3>]
ceph_update_writeable_page+0xe3/0x590 [ceph]
Jun  4 22:19:03 hp1 kernel: [ 7627.132058] PGD 795e3067 PUD 37673067 PMD 0
Jun  4 22:19:03 hp1 kernel: [ 7627.132065] Oops: 0000 [#1] PREEMPT SMP
...
Jun  4 22:19:03 hp1 kernel: [ 7627.132213] Call Trace:
Jun  4 22:19:03 hp1 kernel: [ 7627.132247]  [<ffffffffa01a24ec>]
ceph_write_begin+0x6c/0x100 [ceph]
Jun  4 22:19:03 hp1 kernel: [ 7627.132267]  [<ffffffff810f6652>]
generic_perform_write+0xc2/0x200
Jun  4 22:19:03 hp1 kernel: [ 7627.132277]  [<ffffffff810f67ea>]
generic_file_buffered_write+0x5a/0x90
Jun  4 22:19:03 hp1 kernel: [ 7627.132284]  [<ffffffff810f7dc9>]
__generic_file_aio_write+0x219/0x410
Jun  4 22:19:03 hp1 kernel: [ 7627.132293]  [<ffffffff810f802f>]
generic_file_aio_write+0x6f/0xf0
Jun  4 22:19:03 hp1 kernel: [ 7627.132306]  [<ffffffffa019cfbf>]
ceph_aio_write+0x2cf/0x580 [ceph]
Jun  4 22:19:03 hp1 kernel: [ 7627.132323]  [<ffffffff811533d8>]
do_sync_write+0xb8/0xf0
Jun  4 22:19:03 hp1 kernel: [ 7627.132330]  [<ffffffff81153bce>]
vfs_write+0xae/0x180
Jun  4 22:19:03 hp1 kernel: [ 7627.132337]  [<ffffffff81153ef7>]
sys_write+0x47/0x90
Jun  4 22:19:03 hp1 kernel: [ 7627.132344]  [<ffffffff815a5d12>]
system_call_fastpath+0x16/0x1b

Any takers?

Now, while we're interested in resolving this bug, we are currently
focusing on RADOS, RBD, and radosgw, and that means that the Ceph
Distributed File System is not getting as much of our time. It may
take a while for us to come back. If none of the developers see an
easy fix for this, we'll need to file it in the bug tracker and come
back to it in a few months.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux