Indeed, this is the PVE forum post I saw earlier. /Z On Tue, Oct 12, 2021 at 9:27 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx> wrote: > Igor, > > Thanks for the response. One that I found was: > https://forum.proxmox.com/threads/pve-7-0-bug-kernel-null-pointer-dereference-address-00000000000000c0-pf-error_code-0x0000-no-web-access-no-ssh.96598/ > > In regards to your questions, this is a new cluster deployed at 16.2.6. > > It currently has less than 40TB of data, and is dedicated to CephFS. We > are copying data over from a second Ceph cluster running Nautilus to > mitigate the risk of an in place upgrade. > > > > On Tue, Oct 12, 2021 at 2:22 PM Igor Fedotov <igor.fedotov@xxxxxxxx> > wrote: > >> Zakhar, >> >> could you please point me to the similar reports at Proxmox forum? >> >> Curious what's the Ceph release mentioned there... >> >> Thanks, >> >> Igor >> >> On 10/12/2021 8:53 PM, Zakhar Kirpichenko wrote: >> > Hi, >> > >> > This could be kernel-related, as I've seen similar reports in Proxmox >> > forum. Specifically, 5.11.x with Ceph seems to be hitting kernel NULL >> > pointer dereference. Perhaps a newer kernel would help. If not, I'm >> running >> > 16.2.6 with kernel 5.4.x without any issues. >> > >> > Best regards, >> > Z >> > >> > On Tue, Oct 12, 2021 at 8:31 PM Marco Pizzolo <marcopizzolo@xxxxxxxxx> >> > wrote: >> > >> >> Hello everyone, >> >> >> >> We are seeing instability in 20.04.3 using HWE kernel and Ceph 16.2.6 >> >> w/Podman. >> >> >> >> We have OSDs that fail after <24 hours and I'm not sure why. >> >> >> >> Seeing this: >> >> >> >> ceph crash info >> >> 2021-10-12T14:32:49.169552Z_d1ee94f7-1aaa-4221-abeb-68bd56d3c763 >> >> { >> >> "backtrace": [ >> >> "/lib64/libpthread.so.0(+0x12b20) [0x7f4d31099b20]", >> >> "pthread_cond_wait()", >> >> >> >> "(std::condition_variable::wait(std::unique_lock<std::mutex>&)+0x10) >> >> [0x7f4d306de8f0]", >> >> "(Throttle::_wait(long, std::unique_lock<std::mutex>&)+0x10d) >> >> [0x55c52a0f077d]", >> >> "(Throttle::get(long, long)+0xb9) [0x55c52a0f1199]", >> >> >> "(BlueStore::BlueStoreThrottle::try_start_transaction(KeyValueDB&, >> >> BlueStore::TransContext&, std::chrono::time_point<ceph::mono_clock, >> >> std::chrono::duration<unsigned long, std::ratio<1l, 1000000000l> > >> >)+0x29) >> >> [0x55c529f362c9]", >> >> >> >> >> >> >> "(BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, >> >> std::vector<ceph::os::Transaction, >> std::allocator<ceph::os::Transaction> >> >>> &, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x854) >> >> [0x55c529fb7664]", >> >> "(non-virtual thunk to >> >> PrimaryLogPG::queue_transactions(std::vector<ceph::os::Transaction, >> >> std::allocator<ceph::os::Transaction> >&, >> >> boost::intrusive_ptr<OpRequest>)+0x58) [0x55c529c0ee98]", >> >> "(ReplicatedBackend::submit_transaction(hobject_t const&, >> >> object_stat_sum_t const&, eversion_t const&, >> std::unique_ptr<PGTransaction, >> >> std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t >> >> const&, std::vector<pg_log_entry_t, std::allocator<pg_log_entry_t> >&&, >> >> std::optional<pg_hit_set_history_t>&, Context*, unsigned long, >> osd_reqid_t, >> >> boost::intrusive_ptr<OpRequest>)+0xcad) [0x55c529dfbedd]", >> >> "(PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*, >> >> PrimaryLogPG::OpContext*)+0xcf0) [0x55c529b7a630]", >> >> "(PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0x115d) >> >> [0x55c529bd65ed]", >> >> >> "(PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x2de2) >> >> [0x55c529bdf162]", >> >> "(PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, >> >> ThreadPool::TPHandle&)+0xd1c) [0x55c529be64ac]", >> >> "(OSD::dequeue_op(boost::intrusive_ptr<PG>, >> >> boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x309) >> >> [0x55c529a6f1b9]", >> >> "(ceph::osd::scheduler::PGOpItem::run(OSD*, OSDShard*, >> >> boost::intrusive_ptr<PG>&, ThreadPool::TPHandle&)+0x68) >> [0x55c529ccc868]", >> >> "(OSD::ShardedOpWQ::_process(unsigned int, >> >> ceph::heartbeat_handle_d*)+0xa58) [0x55c529a8f1e8]", >> >> "(ShardedThreadPool::shardedthreadpool_worker(unsigned >> int)+0x5c4) >> >> [0x55c52a0fa6c4]", >> >> "(ShardedThreadPool::WorkThreadSharded::entry()+0x14) >> >> [0x55c52a0fd364]", >> >> "/lib64/libpthread.so.0(+0x814a) [0x7f4d3108f14a]", >> >> "clone()" >> >> ], >> >> "ceph_version": "16.2.6", >> >> "crash_id": >> >> "2021-10-12T14:32:49.169552Z_d1ee94f7-1aaa-4221-abeb-68bd56d3c763", >> >> "entity_name": "osd.14", >> >> "os_id": "centos", >> >> "os_name": "CentOS Linux", >> >> "os_version": "8", >> >> "os_version_id": "8", >> >> "process_name": "ceph-osd", >> >> "stack_sig": >> >> "46b81ca079908da081327cbc114a9c1801dfdbb81303b85fff0d4107a1aeeabe", >> >> "timestamp": "2021-10-12T14:32:49.169552Z", >> >> "utsname_hostname": "<HOSTNAME_REMOVED>", >> >> "utsname_machine": "x86_64", >> >> "utsname_release": "5.11.0-37-generic", >> >> "utsname_sysname": "Linux", >> >> "utsname_version": "#41~20.04.2-Ubuntu SMP Fri Sep 24 09:06:38 UTC >> >> 2021" >> >> >> >> dmesg on host shows: >> >> >> >> [66258.080040] BUG: kernel NULL pointer dereference, address: >> >> 00000000000000c0 >> >> [66258.080067] #PF: supervisor read access in kernel mode >> >> [66258.080081] #PF: error_code(0x0000) - not-present page >> >> [66258.080093] PGD 0 P4D 0 >> >> [66258.080105] Oops: 0000 [#1] SMP NOPTI >> >> [66258.080115] CPU: 35 PID: 4955 Comm: zabbix_agentd Not tainted >> >> 5.11.0-37-generic #41~20.04.2-Ubuntu >> >> [66258.080137] Hardware name: Supermicro SSG-6049P-E1CR60L+/X11DSC+, >> BIOS >> >> 3.3 02/21/2020 >> >> [66258.080154] RIP: 0010:blk_mq_put_rq_ref+0xa/0x60 >> >> [66258.080171] Code: 15 0f b6 d3 4c 89 e7 be 01 00 00 00 e8 cf fe ff >> ff 5b >> >> 41 5c 5d c3 0f 0b 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 55 48 8b 47 10 >> >> <48> 8b 80 c0 00 00 00 48 89 e5 48 3b 78 40 74 1f 4c 8d 87 e8 00 00 >> >> [66258.080210] RSP: 0018:ffffa251249fbbc0 EFLAGS: 00010283 >> >> [66258.080224] RAX: 0000000000000000 RBX: ffffa251249fbc48 RCX: >> >> 0000000000000002 >> >> [66258.080240] RDX: 0000000000000001 RSI: 0000000000000202 RDI: >> >> ffff9382f4d8e000 >> >> [66258.080256] RBP: ffffa251249fbbf8 R08: 0000000000000000 R09: >> >> 0000000000000035 >> >> [66258.080272] R10: abcc77118461cefd R11: ffff93b382257076 R12: >> >> ffff9382f4d8e000 >> >> [66258.080288] R13: ffff9382f5670c00 R14: 0000000000000000 R15: >> >> 0000000000000001 >> >> [66258.080304] FS: 00007fedc0ef46c0(0000) GS:ffff93e13f4c0000(0000) >> >> knlGS:0000000000000000 >> >> [66258.080322] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >> >> [66258.080335] CR2: 00000000000000c0 CR3: 000000011cc44005 CR4: >> >> 00000000007706e0 >> >> [66258.080351] DR0: 0000000000000000 DR1: 0000000000000000 DR2: >> >> 0000000000000000 >> >> [66258.080367] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: >> >> 0000000000000400 >> >> [66258.080383] PKRU: 55555554 >> >> [66258.080391] Call Trace: >> >> [66258.080401] ? bt_iter+0x54/0x90 >> >> [66258.080413] blk_mq_queue_tag_busy_iter+0x18b/0x2d0 >> >> [66258.080427] ? blk_mq_hctx_mark_pending+0x70/0x70 >> >> [66258.080440] ? blk_mq_hctx_mark_pending+0x70/0x70 >> >> [66258.080452] blk_mq_in_flight+0x38/0x60 >> >> [66258.080463] diskstats_show+0x75/0x2b0 >> >> [66258.080475] traverse+0x78/0x200 >> >> [66258.080485] seq_lseek+0x61/0xd0 >> >> [66258.080495] proc_reg_llseek+0x77/0xa0 >> >> [66258.080507] ksys_lseek+0x68/0xb0 >> >> [66258.080519] __x64_sys_lseek+0x1a/0x20 >> >> [66258.080530] do_syscall_64+0x38/0x90 >> >> [66258.080542] entry_SYSCALL_64_after_hwframe+0x44/0xa9 >> >> [66258.080557] RIP: 0033:0x7fedc237293b >> >> [66258.081098] Code: c3 48 8b 15 8f 96 00 00 f7 d8 64 89 02 b8 ff ff >> ff ff >> >> eb be 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa b8 08 00 00 00 0f 05 >> >> <48> 3d 00 f0 ff ff 77 05 c3 0f 1f 40 00 48 8b 15 59 96 00 00 f7 d8 >> >> [66258.082122] RSP: 002b:00007fffb1cfd8e8 EFLAGS: 00000206 ORIG_RAX: >> >> 0000000000000008 >> >> [66258.082633] RAX: ffffffffffffffda RBX: 0000000000000004 RCX: >> >> 00007fedc237293b >> >> [66258.083129] RDX: 0000000000000000 RSI: 00000000000022dc RDI: >> >> 0000000000000007 >> >> [66258.083612] RBP: 00007fffb1cfd950 R08: 0000000000000002 R09: >> >> 0000000061659bf0 >> >> [66258.084080] R10: 0000000000000000 R11: 0000000000000206 R12: >> >> 000055a02d1acd4d >> >> [66258.084536] R13: 0000000000000003 R14: 0000000000000000 R15: >> >> 0000000000000000 >> >> [66258.084979] Modules linked in: binfmt_misc overlay bonding >> nls_iso8859_1 >> >> dm_multipath scsi_dh_rdac scsi_dh_emc scsi_dh_alua intel_rapl_msr >> >> intel_rapl_common isst_if_common skx_edac nfit x86_pkg_temp_thermal >> >> coretemp kvm_intel ipmi_ssif kvm rapl intel_cstate efi_pstore joydev >> >> input_leds intel_pch_thermal mei_me mei ioatdma acpi_ipmi ipmi_si >> >> ipmi_devintf ipmi_msghandler acpi_power_meter mac_hid sch_fq_codel msr >> >> ip_tables x_tables autofs4 btrfs blake2b_generic raid10 raid456 >> >> async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq >> >> libcrc32c raid0 multipath linear mlx5_ib ib_uverbs ib_core hid_generic >> >> usbhid hid ses enclosure scsi_transport_sas raid1 crct10dif_pclmul >> >> crc32_pclmul ast drm_vram_helper i2c_algo_bit drm_ttm_helper >> >> ghash_clmulni_intel ttm mlx5_core drm_kms_helper aesni_intel >> syscopyarea >> >> crypto_simd sysfillrect sysimgblt cryptd glue_helper fb_sys_fops cec >> >> pci_hyperv_intf rc_core mlxfw megaraid_sas tls drm ixgbe xfrm_algo dca >> mdio >> >> ahci i2c_i801 vmd xhci_pci lpc_ich >> >> [66258.085039] libahci i2c_smbus xhci_pci_renesas wmi >> >> [66258.089332] CR2: 00000000000000c0 >> >> [66258.089825] ---[ end trace c1ae715ae7a3e043 ]--- >> >> [66258.144753] RIP: 0010:blk_mq_put_rq_ref+0xa/0x60 >> >> [66258.145285] Code: 15 0f b6 d3 4c 89 e7 be 01 00 00 00 e8 cf fe ff >> ff 5b >> >> 41 5c 5d c3 0f 0b 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 55 48 8b 47 10 >> >> <48> 8b 80 c0 00 00 00 48 89 e5 48 3b 78 40 74 1f 4c 8d 87 e8 00 00 >> >> [66258.146271] RSP: 0018:ffffa251249fbbc0 EFLAGS: 00010283 >> >> [66258.146768] RAX: 0000000000000000 RBX: ffffa251249fbc48 RCX: >> >> 0000000000000002 >> >> [66258.147268] RDX: 0000000000000001 RSI: 0000000000000202 RDI: >> >> ffff9382f4d8e000 >> >> [66258.147779] RBP: ffffa251249fbbf8 R08: 0000000000000000 R09: >> >> 0000000000000035 >> >> [66258.148309] R10: abcc77118461cefd R11: ffff93b382257076 R12: >> >> ffff9382f4d8e000 >> >> [66258.148802] R13: ffff9382f5670c00 R14: 0000000000000000 R15: >> >> 0000000000000001 >> >> [66258.149291] FS: 00007fedc0ef46c0(0000) GS:ffff93e13f4c0000(0000) >> >> knlGS:0000000000000000 >> >> [66258.149785] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >> >> [66258.150281] CR2: 00000000000000c0 CR3: 000000011cc44005 CR4: >> >> 00000000007706e0 >> >> [66258.150791] DR0: 0000000000000000 DR1: 0000000000000000 DR2: >> >> 0000000000000000 >> >> [66258.151290] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: >> >> 0000000000000400 >> >> [66258.151789] PKRU: 55555554 >> >> [69435.862190] perf: interrupt took too long (2513 > 2500), lowering >> >> kernel.perf_event_max_sample_rate to 79500 >> >> >> >> Any guidance is much appreciated. >> >> >> >> Thanks, >> >> Marco >> >> _______________________________________________ >> >> ceph-users mailing list -- ceph-users@xxxxxxx >> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx >> >> >> > _______________________________________________ >> > ceph-users mailing list -- ceph-users@xxxxxxx >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx >> > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx