Re: XFS Calltraces by using XFS with Ceph

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Jun 14, 2017 at 03:22:36PM +0200, list@xxxxxxxxxxxxxxx wrote:
> I get this output of gdb:
> 
> # gdb /usr/lib/debug/lib/modules/4.4.0-75-generic/kernel/fs/xfs/xfs.ko
> GNU gdb (Ubuntu 7.11.1-0ubuntu1~16.04) 7.11.1
> [...]
> Reading symbols from
> /usr/lib/debug/lib/modules/4.4.0-75-generic/kernel/fs/xfs/xfs.ko...done.
> (gdb) list *xfs_da3_node_read+0x30
> 0x2b5d0 is in xfs_da3_node_read
> (/build/linux-Hlembm/linux-4.4.0/fs/xfs/libxfs/xfs_da_btree.c:270).
> 265	/build/linux-Hlembm/linux-4.4.0/fs/xfs/libxfs/xfs_da_btree.c: No such
> file or directory.
> 

This would be more helpful if you had the source code available. :P

Anyways, I suspect you have a NULL buffer (though I'm not sure where the
0xa0 offset comes from). There have been a couple fixes in that area
that come to mind, but it looks to me that v4.4 kernels should already
have them.  Otherwise, this doesn't ring any bells for me. Perhaps
somebody else can chime in on that.

I suppose the best next step is to try a more recent, non-distro kernel.
If the problem still occurs, see if you can provide a crash dump for
analysis.

Brian

> Am 2017-06-14 14:08, schrieb Brian Foster:
> > On Wed, Jun 14, 2017 at 10:22:38AM +0200, list@xxxxxxxxxxxxxxx wrote:
> > > Hello guys,
> > > 
> > > we have currently an issue with our ceph setup based on XFS.
> > > Sometimes some
> > > nodes are dying with high load with this calltrace in dmesg:
> > > 
> > > [Tue Jun 13 13:18:48 2017] BUG: unable to handle kernel NULL pointer
> > > dereference at 00000000000000a0
> > > [Tue Jun 13 13:18:48 2017] IP: [<ffffffffc06555a0>]
> > > xfs_da3_node_read+0x30/0xb0 [xfs]
> > > [Tue Jun 13 13:18:48 2017] PGD 0
> > > [Tue Jun 13 13:18:48 2017] Oops: 0000 [#1] SMP
> > > [Tue Jun 13 13:18:48 2017] Modules linked in: cpuid arc4 md4
> > > nls_utf8 cifs
> > > fscache nfnetlink_queue nfnetlink xt_CHECKSUM xt_nat iptable_nat
> > > nf_nat_ipv4
> > > xt_NFQUEUE xt_CLASSIFY ip6table_mangle dccp_diag dccp tcp_diag
> > > udp_diag
> > > inet_diag unix_diag af_packet_diag netlink_diag veth dummy bridge
> > > stp llc
> > > ebtable_filter ebtables iptable_mangle xt_CT iptable_raw
> > > nf_conntrack_ipv4
> > > nf_defrag_ipv4 iptable_filter ip_tables xt_tcpudp nf_conntrack_ipv6
> > > nf_defrag_ipv6 xt_conntrack ip6table_filter ip6_tables x_tables xfs
> > > ipmi_devintf dcdbas x86_pkg_temp_thermal intel_powerclamp coretemp
> > > crct10dif_pclmul crc32_pclmul ghash_clmulni_intel aesni_intel
> > > ipmi_ssif
> > > aes_x86_64 lrw gf128mul glue_helper ablk_helper cryptd sb_edac
> > > edac_core
> > > input_leds joydev lpc_ich ioatdma shpchp 8250_fintek ipmi_si
> > > ipmi_msghandler
> > > acpi_pad acpi_power_meter
> > > [Tue Jun 13 13:18:48 2017]  mac_hid vhost_net vhost macvtap macvlan
> > > kvm_intel kvm irqbypass cdc_ether nf_nat_ftp tcp_htcp nf_nat_pptp
> > > nf_nat_proto_gre nf_conntrack_ftp bonding nf_nat_sip
> > > nf_conntrack_sip nf_nat
> > > nf_conntrack_pptp nf_conntrack_proto_gre nf_conntrack usbnet mii lp
> > > parport
> > > autofs4 btrfs raid456 async_raid6_recov async_memcpy async_pq
> > > async_xor
> > > async_tx xor raid6_pq libcrc32c raid0 multipath linear raid10 raid1
> > > hid_generic usbhid hid ixgbe igb vxlan ip6_udp_tunnel ahci dca
> > > udp_tunnel
> > > libahci i2c_algo_bit ptp megaraid_sas pps_core mdio wmi fjes
> > > [Tue Jun 13 13:18:48 2017] CPU: 3 PID: 3844 Comm: tp_fstore_op Not
> > > tainted
> > > 4.4.0-75-generic #96-Ubuntu
> > > [Tue Jun 13 13:18:48 2017] Hardware name: Dell Inc. PowerEdge
> > > R720/0XH7F2,
> > > BIOS 2.5.4 01/22/2016
> > > [Tue Jun 13 13:18:48 2017] task: ffff881feda65400 ti: ffff883fbda08000
> > > task.ti: ffff883fbda08000
> > > [Tue Jun 13 13:18:48 2017] RIP: 0010:[<ffffffffc06555a0>]
> > > [<ffffffffc06555a0>] xfs_da3_node_read+0x30/0xb0 [xfs]
> > 
> > What line does this point at (i.e., 'list *xfs_da3_node_read+0x30' from
> > gdb) on your kernel?
> > 
> > Brian
> > 
> > > [Tue Jun 13 13:18:48 2017] RSP: 0018:ffff883fbda0bc88  EFLAGS:
> > > 00010286
> > > [Tue Jun 13 13:18:48 2017] RAX: 0000000000000000 RBX:
> > > ffff8801102c5050 RCX:
> > > 0000000000000001
> > > [Tue Jun 13 13:18:48 2017] RDX: 0000000000000000 RSI:
> > > 0000000000000000 RDI:
> > > ffff883fbda0bc38
> > > [Tue Jun 13 13:18:48 2017] RBP: ffff883fbda0bca8 R08:
> > > 0000000000000001 R09:
> > > fffffffffffffffe
> > > [Tue Jun 13 13:18:48 2017] R10: ffff880007374ae0 R11:
> > > 0000000000000001 R12:
> > > ffff883fbda0bcd8
> > > [Tue Jun 13 13:18:48 2017] R13: ffff880035ac4c80 R14:
> > > 0000000000000001 R15:
> > > 000000008b1f4885
> > > [Tue Jun 13 13:18:48 2017] FS:  00007fc574607700(0000)
> > > GS:ffff883fff040000(0000) knlGS:0000000000000000
> > > [Tue Jun 13 13:18:48 2017] CS:  0010 DS: 0000 ES: 0000 CR0:
> > > 0000000080050033
> > > [Tue Jun 13 13:18:48 2017] CR2: 00000000000000a0 CR3:
> > > 0000003fd828d000 CR4:
> > > 00000000001426e0
> > > [Tue Jun 13 13:18:48 2017] Stack:
> > > [Tue Jun 13 13:18:48 2017]  ffffffffc06b4b50 ffffffffc0695ecc
> > > ffff883fbda0bde0 0000000000000001
> > > [Tue Jun 13 13:18:48 2017]  ffff883fbda0bd20 ffffffffc06718b3
> > > 0000000300000008 ffff880e99b44010
> > > [Tue Jun 13 13:18:48 2017]  00000000360c65a8 ffff88270f80b900
> > > 0000000000000000 0000000000000000
> > > [Tue Jun 13 13:18:48 2017] Call Trace:
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc0695ecc>] ?
> > > xfs_trans_roll+0x2c/0x50
> > > [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc06718b3>]
> > > xfs_attr3_node_inactive+0x183/0x220 [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc06718f9>]
> > > xfs_attr3_node_inactive+0x1c9/0x220 [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc06719fc>]
> > > xfs_attr3_root_inactive+0xac/0x100 [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc0671b9c>]
> > > xfs_attr_inactive+0x14c/0x1a0 [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc068bda5>]
> > > xfs_inactive+0x85/0x120
> > > [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffffc06912f5>]
> > > xfs_fs_evict_inode+0xa5/0x100 [xfs]
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffff8122a90e>] evict+0xbe/0x190
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffff8122abf1>] iput+0x1c1/0x240
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffff8121f6b9>]
> > > do_unlinkat+0x199/0x2d0
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffff81220256>] SyS_unlink+0x16/0x20
> > > [Tue Jun 13 13:18:48 2017]  [<ffffffff8183b972>]
> > > entry_SYSCALL_64_fastpath+0x16/0x71
> > > [Tue Jun 13 13:18:48 2017] Code: 55 48 89 e5 41 54 53 4d 89 c4 48 89
> > > fb 48
> > > 83 ec 10 48 c7 04 24 50 4b 6b c0 e8 dd fe ff ff 85 c0 75 46 48 85 db
> > > 74 41
> > > 49 8b 34 24 <48> 8b 96 a0 00 00 00 0f b7 52 08 66 c1 c2 08 66 81 fa
> > > be 3e 74
> > > [Tue Jun 13 13:18:48 2017] RIP  [<ffffffffc06555a0>]
> > > xfs_da3_node_read+0x30/0xb0 [xfs]
> > > [Tue Jun 13 13:18:48 2017]  RSP <ffff883fbda0bc88>
> > > [Tue Jun 13 13:18:48 2017] CR2: 00000000000000a0
> > > [Tue Jun 13 13:18:48 2017] ---[ end trace 5470d0d55cacb4ef ]---
> > > 
> > > The ceph OSD running on this server has then the issue that it can
> > > not reach
> > > any other osd in the pool.
> > > 
> > >  -1043> 2017-06-13 13:24:00.917597 7fc539a72700  0 --
> > > 192.168.14.19:6827/3389 >> 192.168.14.7:6805/3658
> > > pipe(0x558219846000 sd=23
> > > :6827
> > > s=0 pgs=0 cs=0 l=0 c=0x55821a330400).accept connect_seq 7 vs
> > > existing 7
> > > state standby
> > >  -1042> 2017-06-13 13:24:00.918433 7fc539a72700  0 --
> > > 192.168.14.19:6827/3389 >> 192.168.14.7:6805/3658
> > > pipe(0x558219846000 sd=23
> > > :6827
> > > s=0 pgs=0 cs=0 l=0 c=0x55821a330400).accept connect_seq 8 vs
> > > existing 7
> > > state standby
> > >  -1041> 2017-06-13 13:24:03.654983 7fc4dd21d700  0 --
> > > 192.168.14.19:6825/3389 >> :/0 pipe(0x5581fa6ba000 sd=524 :6825 s=0
> > > pgs=0
> > > cs=0 l=0
> > >  c=0x55820a9e5000).accept failed to getpeername (107) Transport
> > > endpoint is
> > > not connected
> > > 
> > > 
> > > There are a lot more of these messages. Does any of you have the
> > > same issue?
> > > We are running Ubuntu 16.04 with kernel 4.4.0-75.96.
> > > 
> > > Best regards,
> > > Jonas
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe linux-xfs"
> > > in
> > > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [XFS Filesystem Development (older mail)]     [Linux Filesystem Development]     [Linux Audio Users]     [Yosemite Trails]     [Linux Kernel]     [Linux RAID]     [Linux SCSI]


  Powered by Linux