Open zrav opened 2 years ago
Hi #11679, nice to see you again.
@rincebrain if this is a duplicate, feel free to close. In the meantime, I had the last stack trace again...
It may be the most duplicated bug in recent memory. I don't think they've been being closed so far, though, since some of them are more ambiguously duplicates than others.
Another one here with 2.1.9-1~bpo11+1.
[3514154.883087] ------------[ cut here ]------------
[3514154.888230] virt_to_cache: Object is not a Slab page!
[3514154.893808] WARNING: CPU: 7 PID: 3387058 at mm/slab.h:620 kmem_cache_free+0x35d/0x410
[3514154.902133] Modules linked in: xt_nat nf_conntrack_netlink xfrm_user binfmt_misc veth tun xt_CHECKSUM xt_MASQUERADE xt_tcpudp nft_compat overlay macvlan bridge stp llc nls_ascii nls_cp437 vfat fat nft_nat nft_masq nft_fib_ipv4 nft_fib nft_chain_nat nf_nat nft_reject_inet nf_reject_ipv4 nf_reject_ipv6 nft_reject nft_limit intel_rapl_msr intel_rapl_common ipmi_ssif x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm nft_ct irqbypass nf_conntrack ast nf_defrag_ipv6 ghash_clmulni_intel nf_defrag_ipv4 drm_vram_helper acpi_ipmi aesni_intel drm_ttm_helper cdc_ether iTCO_wdt crypto_simd ttm cryptd intel_pmc_bxt usbnet ipmi_si mei_me intel_cstate iTCO_vendor_support drm_kms_helper joydev pcspkr evdev ipmi_devintf intel_uncore efi_pstore mii i2c_algo_bit mei watchdog sg ipmi_msghandler acpi_pad intel_pmc_core acpi_power_meter button acpi_tad nf_tables crc32c_generic nfnetlink fuse drm configfs ip_tables x_tables autofs4 zfs(POE) zunicode(POE) zzstd(OE) zlua(OE) zavl(POE) icp(POE)
[3514154.902177] zcommon(POE) znvpair(POE) spl(OE) efivarfs raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq libcrc32c raid1 raid0 multipath linear md_mod sr_mod cdrom hid_generic usbhid hid uas usb_storage sd_mod ixgbe xfrm_algo ahci xhci_pci dca libahci nvme mdio_devres xhci_hcd libata nvme_core libphy t10_pi intel_ish_ipc intel_lpss_pci i2c_i801 crc64_rocksoft crc32_pclmul ptp scsi_mod crc64 usbcore intel_lpss crc32c_intel i2c_smbus intel_ishtp pps_core crc_t10dif idma64 mdio usb_common crct10dif_generic scsi_common crct10dif_pclmul crct10dif_common fan video
[3514155.046131] CPU: 7 PID: 3387058 Comm: receive_writer Tainted: P OE 6.0.0-0.deb11.6-amd64 #1 Debian 6.0.12-1~bpo11+1
[3514155.058687] Hardware name: GIGABYTE MX33-BS1-V1/MX33-BS1-V1, BIOS F04a 10/01/2021
[3514155.066642] RIP: 0010:kmem_cache_free+0x35d/0x410
[3514155.071823] Code: fd ff ff 80 3d d0 4f 6f 01 00 0f 85 d7 fe ff ff 48 c7 c6 00 de a3 ba 48 c7 c7 48 30 d6 ba c6 05 b5 4f 6f 01 01 e8 29 c9 67 00 <0f> 0b 5b 5d 41 5c 41 5d 41 5e c3 cc cc cc cc 48 8b 57 60 48 8b 4d
[3514155.091324] RSP: 0018:ffffac7894ee3a30 EFLAGS: 00010286
[3514155.097026] RAX: 0000000000000000 RBX: ffff93be21f74000 RCX: 0000000000000000
[3514155.104658] RDX: 0000000000000001 RSI: ffffffffbad7c802 RDI: 00000000ffffffff
[3514155.112293] RBP: ffff93baf07a6400 R08: 0000000000000000 R09: 00000001000005a4
[3514155.119896] R10: ffffac7894ee38d0 R11: ffffffffbb4c2a20 R12: ffff93bea1f74000
[3514155.127489] R13: ffffffffc0d838a0 R14: 0000000000000000 R15: ffffffffc0d838c0
[3514155.135082] FS: 0000000000000000(0000) GS:ffff93c9dfbc0000(0000) knlGS:0000000000000000
[3514155.143617] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[3514155.149814] CR2: 00005595f4b1eac0 CR3: 00000001ac810001 CR4: 0000000000772ee0
[3514155.157405] PKRU: 55555554
[3514155.160553] Call Trace:
[3514155.163446] <TASK>
[3514155.165983] spl_kmem_cache_free+0x12e/0x1d0 [spl]
[3514155.171212] arc_buf_fill+0x3f5/0xce0 [zfs]
[3514155.175867] ? dbuf_dirty+0x498/0x9c0 [zfs]
[3514155.180520] arc_untransform+0x1d/0x80 [zfs]
[3514155.185261] dbuf_read_verify_dnode_crypt+0xf2/0x160 [zfs]
[3514155.191191] dbuf_read_impl.constprop.0+0x52a/0x6d0 [zfs]
[3514155.197027] ? percpu_counter_add_batch+0x58/0xc0
[3514155.202150] ? dbuf_cons+0xa7/0xc0 [zfs]
[3514155.206513] ? spl_kmem_cache_alloc+0x9f/0x7d0 [spl]
[3514155.211887] ? dbuf_rele_and_unlock+0x132/0x690 [zfs]
[3514155.217361] ? preempt_count_add+0x70/0xa0
[3514155.221853] ? preempt_count_add+0x70/0xa0
[3514155.226345] ? _raw_spin_lock+0x13/0x40
[3514155.230570] ? aggsum_add+0x178/0x190 [zfs]
[3514155.235160] dbuf_read+0xe5/0x5f0 [zfs]
[3514155.239402] ? dnode_hold_impl+0x4eb/0xee0 [zfs]
[3514155.244439] ? preempt_count_add+0x70/0xa0
[3514155.248921] dmu_bonus_hold_by_dnode+0x86/0x1a0 [zfs]
[3514155.254376] receive_object+0x410/0xca0 [zfs]
[3514155.259146] ? dnode_rele_and_unlock+0x59/0xf0 [zfs]
[3514155.264508] ? preempt_count_add+0x70/0xa0
[3514155.268976] ? __thread_exit+0x20/0x20 [spl]
[3514155.273616] receive_writer_thread+0x1cc/0xad0 [zfs]
[3514155.278972] ? __slab_free+0xa1/0x2d0
[3514155.283000] ? set_next_task_fair+0x2d/0xd0
[3514155.287533] ? _raw_spin_unlock_irqrestore+0x23/0x40
[3514155.292849] ? receive_process_write_record+0x1a0/0x1a0 [zfs]
[3514155.298972] ? __thread_exit+0x20/0x20 [spl]
[3514155.303669] ? thread_generic_wrapper+0x56/0x70 [spl]
[3514155.309066] thread_generic_wrapper+0x56/0x70 [spl]
[3514155.314273] kthread+0xe5/0x110
[3514155.317761] ? kthread_complete_and_exit+0x20/0x20
[3514155.322885] ret_from_fork+0x1f/0x30
[3514155.326789] </TASK>
[3514155.329331] ---[ end trace 0000000000000000 ]---
System information
Describe the problem you're observing
While receiving an unencrypted dataset that was then compressed and encrypted, the system paniced.
Describe how to reproduce the problem
I don't have a surefire way to reproduce. But it did happen twice in a matter of minutes. Send options -L -p, receive options -x encryption -x compression.
Include any warning/errors/backtraces from the system logs
The pool was hard-locked after that but I was able to do a regular reboot. Shortly after I got the next panic while receiving another dataset:
This time the pool was not frozen, but I rebooted just in case.
Receiving Pool stats: