Re: OSD Crashes in 16.2.6

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Igor,

Thanks for the response.  One that I found was:
https://forum.proxmox.com/threads/pve-7-0-bug-kernel-null-pointer-dereference-address-00000000000000c0-pf-error_code-0x0000-no-web-access-no-ssh.96598/

In regards to your questions, this is a new cluster deployed at 16.2.6.

It currently has less than 40TB of data, and is dedicated to CephFS.  We
are copying data over from a second Ceph cluster running Nautilus to
mitigate the risk of an in place upgrade.



On Tue, Oct 12, 2021 at 2:22 PM Igor Fedotov <igor.fedotov@xxxxxxxx> wrote:

> Zakhar,
>
> could you please point me to the similar reports at Proxmox forum?
>
> Curious what's the Ceph release mentioned there...
>
> Thanks,
>
> Igor
>
> On 10/12/2021 8:53 PM, Zakhar Kirpichenko wrote:
> > Hi,
> >
> > This could be kernel-related, as I've seen similar reports in Proxmox
> > forum. Specifically, 5.11.x with Ceph seems to be hitting kernel NULL
> > pointer dereference. Perhaps a newer kernel would help. If not, I'm
> running
> > 16.2.6 with kernel 5.4.x without any issues.
> >
> > Best regards,
> > Z
> >
> > On Tue, Oct 12, 2021 at 8:31 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx>
> > wrote:
> >
> >> Hello everyone,
> >>
> >> We are seeing instability in 20.04.3 using HWE kernel and Ceph 16.2.6
> >> w/Podman.
> >>
> >> We have OSDs that fail after <24 hours and I'm not sure why.
> >>
> >> Seeing this:
> >>
> >> ceph crash info
> >> 2021-10-12T14:32:49.169552Z_d1ee94f7-1aaa-4221-abeb-68bd56d3c763
> >> {
> >>      "backtrace": [
> >>          "/lib64/libpthread.so.0(+0x12b20) [0x7f4d31099b20]",
> >>          "pthread_cond_wait()",
> >>
> >> "(std::condition_variable::wait(std::unique_lock<std::mutex>&)+0x10)
> >> [0x7f4d306de8f0]",
> >>          "(Throttle::_wait(long, std::unique_lock<std::mutex>&)+0x10d)
> >> [0x55c52a0f077d]",
> >>          "(Throttle::get(long, long)+0xb9) [0x55c52a0f1199]",
> >>
> "(BlueStore::BlueStoreThrottle::try_start_transaction(KeyValueDB&,
> >> BlueStore::TransContext&, std::chrono::time_point<ceph::mono_clock,
> >> std::chrono::duration<unsigned long, std::ratio<1l, 1000000000l> >
> >)+0x29)
> >> [0x55c529f362c9]",
> >>
> >>
> >>
> "(BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&,
> >> std::vector<ceph::os::Transaction, std::allocator<ceph::os::Transaction>
> >>> &, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x854)
> >> [0x55c529fb7664]",
> >>          "(non-virtual thunk to
> >> PrimaryLogPG::queue_transactions(std::vector<ceph::os::Transaction,
> >> std::allocator<ceph::os::Transaction> >&,
> >> boost::intrusive_ptr<OpRequest>)+0x58) [0x55c529c0ee98]",
> >>          "(ReplicatedBackend::submit_transaction(hobject_t const&,
> >> object_stat_sum_t const&, eversion_t const&,
> std::unique_ptr<PGTransaction,
> >> std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t
> >> const&, std::vector<pg_log_entry_t, std::allocator<pg_log_entry_t> >&&,
> >> std::optional<pg_hit_set_history_t>&, Context*, unsigned long,
> osd_reqid_t,
> >> boost::intrusive_ptr<OpRequest>)+0xcad) [0x55c529dfbedd]",
> >>          "(PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*,
> >> PrimaryLogPG::OpContext*)+0xcf0) [0x55c529b7a630]",
> >>          "(PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0x115d)
> >> [0x55c529bd65ed]",
> >>          "(PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x2de2)
> >> [0x55c529bdf162]",
> >>          "(PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&,
> >> ThreadPool::TPHandle&)+0xd1c) [0x55c529be64ac]",
> >>          "(OSD::dequeue_op(boost::intrusive_ptr<PG>,
> >> boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x309)
> >> [0x55c529a6f1b9]",
> >>          "(ceph::osd::scheduler::PGOpItem::run(OSD*, OSDShard*,
> >> boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&)+0x68)
> [0x55c529ccc868]",
> >>          "(OSD::ShardedOpWQ::_process(unsigned int,
> >> ceph::heartbeat_handle_d*)+0xa58) [0x55c529a8f1e8]",
> >>          "(ShardedThreadPool::shardedthreadpool_worker(unsigned
> int)+0x5c4)
> >> [0x55c52a0fa6c4]",
> >>          "(ShardedThreadPool::WorkThreadSharded::entry()+0x14)
> >> [0x55c52a0fd364]",
> >>          "/lib64/libpthread.so.0(+0x814a) [0x7f4d3108f14a]",
> >>          "clone()"
> >>      ],
> >>      "ceph_version": "16.2.6",
> >>      "crash_id":
> >> "2021-10-12T14:32:49.169552Z_d1ee94f7-1aaa-4221-abeb-68bd56d3c763",
> >>      "entity_name": "osd.14",
> >>      "os_id": "centos",
> >>      "os_name": "CentOS Linux",
> >>      "os_version": "8",
> >>      "os_version_id": "8",
> >>      "process_name": "ceph-osd",
> >>      "stack_sig":
> >> "46b81ca079908da081327cbc114a9c1801dfdbb81303b85fff0d4107a1aeeabe",
> >>      "timestamp": "2021-10-12T14:32:49.169552Z",
> >>      "utsname_hostname": "<HOSTNAME_REMOVED>",
> >>      "utsname_machine": "x86_64",
> >>      "utsname_release": "5.11.0-37-generic",
> >>      "utsname_sysname": "Linux",
> >>      "utsname_version": "#41~20.04.2-Ubuntu SMP Fri Sep 24 09:06:38 UTC
> >> 2021"
> >>
> >> dmesg on host shows:
> >>
> >> [66258.080040] BUG: kernel NULL pointer dereference, address:
> >> 00000000000000c0
> >> [66258.080067] #PF: supervisor read access in kernel mode
> >> [66258.080081] #PF: error_code(0x0000) - not-present page
> >> [66258.080093] PGD 0 P4D 0
> >> [66258.080105] Oops: 0000 [#1] SMP NOPTI
> >> [66258.080115] CPU: 35 PID: 4955 Comm: zabbix_agentd Not tainted
> >> 5.11.0-37-generic #41~20.04.2-Ubuntu
> >> [66258.080137] Hardware name: Supermicro SSG-6049P-E1CR60L+/X11DSC+,
> BIOS
> >> 3.3 02/21/2020
> >> [66258.080154] RIP: 0010:blk_mq_put_rq_ref+0xa/0x60
> >> [66258.080171] Code: 15 0f b6 d3 4c 89 e7 be 01 00 00 00 e8 cf fe ff ff
> 5b
> >> 41 5c 5d c3 0f 0b 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 55 48 8b 47 10
> >> <48> 8b 80 c0 00 00 00 48 89 e5 48 3b 78 40 74 1f 4c 8d 87 e8 00 00
> >> [66258.080210] RSP: 0018:ffffa251249fbbc0 EFLAGS: 00010283
> >> [66258.080224] RAX: 0000000000000000 RBX: ffffa251249fbc48 RCX:
> >> 0000000000000002
> >> [66258.080240] RDX: 0000000000000001 RSI: 0000000000000202 RDI:
> >> ffff9382f4d8e000
> >> [66258.080256] RBP: ffffa251249fbbf8 R08: 0000000000000000 R09:
> >> 0000000000000035
> >> [66258.080272] R10: abcc77118461cefd R11: ffff93b382257076 R12:
> >> ffff9382f4d8e000
> >> [66258.080288] R13: ffff9382f5670c00 R14: 0000000000000000 R15:
> >> 0000000000000001
> >> [66258.080304] FS:  00007fedc0ef46c0(0000) GS:ffff93e13f4c0000(0000)
> >> knlGS:0000000000000000
> >> [66258.080322] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> >> [66258.080335] CR2: 00000000000000c0 CR3: 000000011cc44005 CR4:
> >> 00000000007706e0
> >> [66258.080351] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> >> 0000000000000000
> >> [66258.080367] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7:
> >> 0000000000000400
> >> [66258.080383] PKRU: 55555554
> >> [66258.080391] Call Trace:
> >> [66258.080401]  ? bt_iter+0x54/0x90
> >> [66258.080413]  blk_mq_queue_tag_busy_iter+0x18b/0x2d0
> >> [66258.080427]  ? blk_mq_hctx_mark_pending+0x70/0x70
> >> [66258.080440]  ? blk_mq_hctx_mark_pending+0x70/0x70
> >> [66258.080452]  blk_mq_in_flight+0x38/0x60
> >> [66258.080463]  diskstats_show+0x75/0x2b0
> >> [66258.080475]  traverse+0x78/0x200
> >> [66258.080485]  seq_lseek+0x61/0xd0
> >> [66258.080495]  proc_reg_llseek+0x77/0xa0
> >> [66258.080507]  ksys_lseek+0x68/0xb0
> >> [66258.080519]  __x64_sys_lseek+0x1a/0x20
> >> [66258.080530]  do_syscall_64+0x38/0x90
> >> [66258.080542]  entry_SYSCALL_64_after_hwframe+0x44/0xa9
> >> [66258.080557] RIP: 0033:0x7fedc237293b
> >> [66258.081098] Code: c3 48 8b 15 8f 96 00 00 f7 d8 64 89 02 b8 ff ff ff
> ff
> >> eb be 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa b8 08 00 00 00 0f 05
> >> <48> 3d 00 f0 ff ff 77 05 c3 0f 1f 40 00 48 8b 15 59 96 00 00 f7 d8
> >> [66258.082122] RSP: 002b:00007fffb1cfd8e8 EFLAGS: 00000206 ORIG_RAX:
> >> 0000000000000008
> >> [66258.082633] RAX: ffffffffffffffda RBX: 0000000000000004 RCX:
> >> 00007fedc237293b
> >> [66258.083129] RDX: 0000000000000000 RSI: 00000000000022dc RDI:
> >> 0000000000000007
> >> [66258.083612] RBP: 00007fffb1cfd950 R08: 0000000000000002 R09:
> >> 0000000061659bf0
> >> [66258.084080] R10: 0000000000000000 R11: 0000000000000206 R12:
> >> 000055a02d1acd4d
> >> [66258.084536] R13: 0000000000000003 R14: 0000000000000000 R15:
> >> 0000000000000000
> >> [66258.084979] Modules linked in: binfmt_misc overlay bonding
> nls_iso8859_1
> >> dm_multipath scsi_dh_rdac scsi_dh_emc scsi_dh_alua intel_rapl_msr
> >> intel_rapl_common isst_if_common skx_edac nfit x86_pkg_temp_thermal
> >> coretemp kvm_intel ipmi_ssif kvm rapl intel_cstate efi_pstore joydev
> >> input_leds intel_pch_thermal mei_me mei ioatdma acpi_ipmi ipmi_si
> >> ipmi_devintf ipmi_msghandler acpi_power_meter mac_hid sch_fq_codel msr
> >> ip_tables x_tables autofs4 btrfs blake2b_generic raid10 raid456
> >> async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq
> >> libcrc32c raid0 multipath linear mlx5_ib ib_uverbs ib_core hid_generic
> >> usbhid hid ses enclosure scsi_transport_sas raid1 crct10dif_pclmul
> >> crc32_pclmul ast drm_vram_helper i2c_algo_bit drm_ttm_helper
> >> ghash_clmulni_intel ttm mlx5_core drm_kms_helper aesni_intel syscopyarea
> >> crypto_simd sysfillrect sysimgblt cryptd glue_helper fb_sys_fops cec
> >> pci_hyperv_intf rc_core mlxfw megaraid_sas tls drm ixgbe xfrm_algo dca
> mdio
> >> ahci i2c_i801 vmd xhci_pci lpc_ich
> >> [66258.085039]  libahci i2c_smbus xhci_pci_renesas wmi
> >> [66258.089332] CR2: 00000000000000c0
> >> [66258.089825] ---[ end trace c1ae715ae7a3e043 ]---
> >> [66258.144753] RIP: 0010:blk_mq_put_rq_ref+0xa/0x60
> >> [66258.145285] Code: 15 0f b6 d3 4c 89 e7 be 01 00 00 00 e8 cf fe ff ff
> 5b
> >> 41 5c 5d c3 0f 0b 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 55 48 8b 47 10
> >> <48> 8b 80 c0 00 00 00 48 89 e5 48 3b 78 40 74 1f 4c 8d 87 e8 00 00
> >> [66258.146271] RSP: 0018:ffffa251249fbbc0 EFLAGS: 00010283
> >> [66258.146768] RAX: 0000000000000000 RBX: ffffa251249fbc48 RCX:
> >> 0000000000000002
> >> [66258.147268] RDX: 0000000000000001 RSI: 0000000000000202 RDI:
> >> ffff9382f4d8e000
> >> [66258.147779] RBP: ffffa251249fbbf8 R08: 0000000000000000 R09:
> >> 0000000000000035
> >> [66258.148309] R10: abcc77118461cefd R11: ffff93b382257076 R12:
> >> ffff9382f4d8e000
> >> [66258.148802] R13: ffff9382f5670c00 R14: 0000000000000000 R15:
> >> 0000000000000001
> >> [66258.149291] FS:  00007fedc0ef46c0(0000) GS:ffff93e13f4c0000(0000)
> >> knlGS:0000000000000000
> >> [66258.149785] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> >> [66258.150281] CR2: 00000000000000c0 CR3: 000000011cc44005 CR4:
> >> 00000000007706e0
> >> [66258.150791] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> >> 0000000000000000
> >> [66258.151290] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7:
> >> 0000000000000400
> >> [66258.151789] PKRU: 55555554
> >> [69435.862190] perf: interrupt took too long (2513 > 2500), lowering
> >> kernel.perf_event_max_sample_rate to 79500
> >>
> >> Any guidance is much appreciated.
> >>
> >> Thanks,
> >> Marco
> >> _______________________________________________
> >> ceph-users mailing list -- ceph-users@xxxxxxx
> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux