Open 512yottabytes opened 2 years ago
Just tried to cat
the file I've created yesterday using zfs send pool1/dataset1 > /home/user/file.zfs
:
cat /home/user/file.zfs | zfs recv -v -u -s pool2/dataset2
and also got panic:
[ 3955.769256] Process accounting resumed
[ 7336.787010] VERIFY3(sa.sa_magic == SA_MAGIC) failed (2531917562 == 3100762)
[ 7336.787020] PANIC at zfs_quota.c:89:zpl_get_file_info()
[ 7336.787024] Showing stack for process 19920
[ 7336.787028] CPU: 1 PID: 19920 Comm: dp_sync_taskq Tainted: P IOE 5.15.0-48-generic #54-Ubuntu
[ 7336.787033] Hardware name: Hewlett-Packard HP Z600 Workstation/0AE8h, BIOS 786G4 v03.60 02/24/2016
... And so on...
Also have errors in rsync
and same panic trying to:
zfs mount pool1/dataset1 # (to /mnt/1)
zfs mount pool2/dataset2 # (to /mnt/2)
rsync -avpxHES /mnt/1/. /mnt/2/.
zpool scrub pool1
and zpool scrub pool2
finished successfully without any error.
It looks like something corrupted a metadata buffer before it was checksummed and written to disk.
@512yottabytes Would you provide the backtrace that you get when this happens with rsync
?
Hello, @ryao
dmesg
:
[ 359.207437] Process accounting resumed
[ 1569.423938] perf: interrupt took too long (2518 > 2500), lowering kernel.perf_event_max_sample_rate to 79250
[ 1755.729478] perf: interrupt took too long (3428 > 3147), lowering kernel.perf_event_max_sample_rate to 58250
[ 1952.858845] perf: interrupt took too long (4307 > 4285), lowering kernel.perf_event_max_sample_rate to 46250
[ 2157.058386] perf: interrupt took too long (5403 > 5383), lowering kernel.perf_event_max_sample_rate to 37000
[ 2402.056405] perf: interrupt took too long (6845 > 6753), lowering kernel.perf_event_max_sample_rate to 29000
[ 3015.416455] perf: interrupt took too long (8892 > 8556), lowering kernel.perf_event_max_sample_rate to 22250
[ 3356.060044] perf: interrupt took too long (11189 > 11115), lowering kernel.perf_event_max_sample_rate to 17750
[ 3960.051537] Process accounting resumed
[10079.546729] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
[10079.546740] PANIC at zfs_znode.c:339:zfs_znode_sa_init()
[10079.546743] Showing stack for process 847563
[10079.546747] CPU: 9 PID: 847563 Comm: rsync Tainted: P IOE 5.15.0-48-generic #54-Ubuntu
[10079.546751] Hardware name: Hewlett-Packard HP Z600 Workstation/0AE8h, BIOS 786G4 v03.60 02/24/2016
[10079.546754] Call Trace:
[10079.546758] <TASK>
[10079.546762] show_stack+0x52/0x5c
[10079.546771] dump_stack_lvl+0x4a/0x63
[10079.546781] dump_stack+0x10/0x16
[10079.546785] spl_dumpstack+0x29/0x2f [spl]
[10079.546806] spl_panic+0xd1/0xe9 [spl]
[10079.546826] ? __cond_resched+0x1a/0x50
[10079.546832] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10079.546999] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10079.547148] ? dmu_buf_replace_user+0x65/0x80 [zfs]
[10079.547300] ? dmu_buf_set_user+0x13/0x20 [zfs]
[10079.547452] ? dmu_buf_set_user_ie+0x15/0x20 [zfs]
[10079.547604] zfs_znode_sa_init+0xdd/0xf0 [zfs]
[10079.547834] zfs_znode_alloc+0x101/0x5d0 [zfs]
[10079.548063] ? arc_space_consume+0x54/0x100 [zfs]
[10079.548203] ? dbuf_read_bonus+0x102/0x110 [zfs]
[10079.548353] ? RW_WRITE_HELD+0x9/0x30 [zfs]
[10079.548502] ? dmu_buf_unlock_parent+0x65/0xa0 [zfs]
[10079.548654] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10079.548803] ? dbuf_read_impl.constprop.0+0x352/0x490 [zfs]
[10079.548953] ? dbuf_rele_and_unlock+0x134/0x540 [zfs]
[10079.549103] ? __cond_resched+0x1a/0x50
[10079.549107] ? __cond_resched+0x1a/0x50
[10079.549111] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10079.549248] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10079.549385] ? __cond_resched+0x1a/0x50
[10079.549389] ? down_read+0x13/0xa0
[10079.549394] ? wmsum_add+0xe/0x20 [zfs]
[10079.549543] ? dbuf_read+0x1e6/0x5b0 [zfs]
[10079.549693] ? __cond_resched+0x1a/0x50
[10079.549697] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10079.549865] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10079.550033] ? dnode_rele_and_unlock+0x69/0xf0 [zfs]
[10079.550202] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10079.550358] ? RW_WRITE_HELD+0x9/0x30 [zfs]
[10079.550513] ? dmu_object_info_from_dnode+0x92/0xa0 [zfs]
[10079.550671] zfs_zget+0x239/0x280 [zfs]
[10079.550900] zfs_dirent_lock+0x424/0x570 [zfs]
[10079.551129] zfs_dirlook+0x91/0x2d0 [zfs]
[10079.551357] zfs_lookup+0x25e/0x410 [zfs]
[10079.551586] zpl_lookup+0xcf/0x230 [zfs]
[10079.551814] ? capable_wrt_inode_uidgid+0x5f/0x70
[10079.551822] __lookup_slow+0x88/0x150
[10079.551829] walk_component+0x145/0x1c0
[10079.551833] ? zfs_dirty_inode+0x2af/0x330 [zfs]
[10079.552061] path_lookupat+0x6e/0x1c0
[10079.552066] filename_lookup+0xcf/0x1d0
[10079.552071] ? __check_object_size+0x1d/0x30
[10079.552077] ? strncpy_from_user+0x44/0x150
[10079.552084] ? getname_flags.part.0+0x4c/0x1b0
[10079.552089] user_path_at_empty+0x3f/0x60
[10079.552093] vfs_statx+0x7a/0x130
[10079.552099] __do_sys_newfstatat+0x36/0x70
[10079.552103] ? exit_to_user_mode_prepare+0x96/0xb0
[10079.552111] ? syscall_exit_to_user_mode+0x27/0x50
[10079.552118] ? do_syscall_64+0x69/0xc0
[10079.552121] ? exit_to_user_mode_prepare+0x37/0xb0
[10079.552126] ? syscall_exit_to_user_mode+0x27/0x50
[10079.552130] ? __x64_sys_newfstatat+0x1c/0x30
[10079.552134] ? do_syscall_64+0x69/0xc0
[10079.552137] __x64_sys_newfstatat+0x1c/0x30
[10079.552142] do_syscall_64+0x5c/0xc0
[10079.552145] ? do_syscall_64+0x69/0xc0
[10079.552148] ? syscall_exit_to_user_mode+0x27/0x50
[10079.552152] ? __x64_sys_close+0x11/0x50
[10079.552157] ? do_syscall_64+0x69/0xc0
[10079.552160] entry_SYSCALL_64_after_hwframe+0x61/0xcb
[10079.552166] RIP: 0033:0x7fbbb12afeee
[10079.552171] Code: 48 89 f2 b9 00 01 00 00 48 89 fe bf 9c ff ff ff e9 07 00 00 00 0f 1f 80 00 00 00 00 f3 0f 1e fa 41 89 ca b8 06 01 00 00 0f 05 <3d> 00 f0 ff ff 77 0b 31 c0 c3 0f 1f 84 00 00 00 00 00 48 8b 15 09
[10079.552174] RSP: 002b:00007ffc2be9a998 EFLAGS: 00000246 ORIG_RAX: 0000000000000106
[10079.552179] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fbbb12afeee
[10079.552182] RDX: 00007ffc2be9aaa0 RSI: 00007ffc2be9bb30 RDI: 00000000ffffff9c
[10079.552185] RBP: 00007ffc2be9aaa0 R08: 00007ffc2be9bb2f R09: 0000000000000000
[10079.552187] R10: 0000000000000100 R11: 0000000000000246 R12: 0000000000000000
[10079.552190] R13: 00007ffc2be9bb30 R14: 000055776e8adfb0 R15: 00007ffc2be9aaa0
[10079.552194] </TASK>
[10271.494666] INFO: task rsync:847563 blocked for more than 120 seconds.
[10271.494676] Tainted: P IOE 5.15.0-48-generic #54-Ubuntu
[10271.494679] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[10271.494682] task:rsync state:D stack: 0 pid:847563 ppid: 46824 flags:0x00004002
[10271.494689] Call Trace:
[10271.494692] <TASK>
[10271.494696] __schedule+0x23d/0x5a0
[10271.494708] schedule+0x4e/0xc0
[10271.494712] spl_panic+0xe7/0xe9 [spl]
[10271.494737] ? __cond_resched+0x1a/0x50
[10271.494741] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10271.494915] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10271.495064] ? dmu_buf_replace_user+0x65/0x80 [zfs]
[10271.495216] ? dmu_buf_set_user+0x13/0x20 [zfs]
[10271.495368] ? dmu_buf_set_user_ie+0x15/0x20 [zfs]
[10271.495521] zfs_znode_sa_init+0xdd/0xf0 [zfs]
[10271.495750] zfs_znode_alloc+0x101/0x5d0 [zfs]
[10271.495979] ? arc_space_consume+0x54/0x100 [zfs]
[10271.496119] ? dbuf_read_bonus+0x102/0x110 [zfs]
[10271.496269] ? RW_WRITE_HELD+0x9/0x30 [zfs]
[10271.496418] ? dmu_buf_unlock_parent+0x65/0xa0 [zfs]
[10271.496570] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10271.496719] ? dbuf_read_impl.constprop.0+0x352/0x490 [zfs]
[10271.496869] ? dbuf_rele_and_unlock+0x134/0x540 [zfs]
[10271.497019] ? __cond_resched+0x1a/0x50
[10271.497024] ? __cond_resched+0x1a/0x50
[10271.497027] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10271.497165] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10271.497302] ? __cond_resched+0x1a/0x50
[10271.497306] ? down_read+0x13/0xa0
[10271.497311] ? wmsum_add+0xe/0x20 [zfs]
[10271.497460] ? dbuf_read+0x1e6/0x5b0 [zfs]
[10271.497611] ? __cond_resched+0x1a/0x50
[10271.497615] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10271.497782] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10271.497950] ? dnode_rele_and_unlock+0x69/0xf0 [zfs]
[10271.498119] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10271.498276] ? RW_WRITE_HELD+0x9/0x30 [zfs]
[10271.498461] ? dmu_object_info_from_dnode+0x92/0xa0 [zfs]
[10271.498703] zfs_zget+0x239/0x280 [zfs]
[10271.498840] zfs_dirent_lock+0x424/0x570 [zfs]
[10271.498978] zfs_dirlook+0x91/0x2d0 [zfs]
[10271.499206] zfs_lookup+0x25e/0x410 [zfs]
[10271.499435] zpl_lookup+0xcf/0x230 [zfs]
[10271.499663] ? capable_wrt_inode_uidgid+0x5f/0x70
[10271.499670] __lookup_slow+0x88/0x150
[10271.499676] walk_component+0x145/0x1c0
[10271.499680] ? zfs_dirty_inode+0x2af/0x330 [zfs]
[10271.499909] path_lookupat+0x6e/0x1c0
[10271.499913] filename_lookup+0xcf/0x1d0
[10271.499919] ? __check_object_size+0x1d/0x30
[10271.499924] ? strncpy_from_user+0x44/0x150
[10271.499930] ? getname_flags.part.0+0x4c/0x1b0
[10271.499935] user_path_at_empty+0x3f/0x60
[10271.499940] vfs_statx+0x7a/0x130
[10271.499946] __do_sys_newfstatat+0x36/0x70
[10271.499950] ? exit_to_user_mode_prepare+0x96/0xb0
[10271.499957] ? syscall_exit_to_user_mode+0x27/0x50
[10271.499963] ? do_syscall_64+0x69/0xc0
[10271.499966] ? exit_to_user_mode_prepare+0x37/0xb0
[10271.499971] ? syscall_exit_to_user_mode+0x27/0x50
[10271.499975] ? __x64_sys_newfstatat+0x1c/0x30
[10271.499980] ? do_syscall_64+0x69/0xc0
[10271.499983] __x64_sys_newfstatat+0x1c/0x30
[10271.499987] do_syscall_64+0x5c/0xc0
[10271.499991] ? do_syscall_64+0x69/0xc0
[10271.499994] ? syscall_exit_to_user_mode+0x27/0x50
[10271.499998] ? __x64_sys_close+0x11/0x50
[10271.500003] ? do_syscall_64+0x69/0xc0
[10271.500006] entry_SYSCALL_64_after_hwframe+0x61/0xcb
[10271.500012] RIP: 0033:0x7fbbb12afeee
[10271.500016] RSP: 002b:00007ffc2be9a998 EFLAGS: 00000246 ORIG_RAX: 0000000000000106
[10271.500021] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fbbb12afeee
[10271.500024] RDX: 00007ffc2be9aaa0 RSI: 00007ffc2be9bb30 RDI: 00000000ffffff9c
[10271.500027] RBP: 00007ffc2be9aaa0 R08: 00007ffc2be9bb2f R09: 0000000000000000
[10271.500029] R10: 0000000000000100 R11: 0000000000000246 R12: 0000000000000000
[10271.500032] R13: 00007ffc2be9bb30 R14: 000055776e8adfb0 R15: 00007ffc2be9aaa0
[10271.500036] </TASK>
[10392.326030] INFO: task rsync:847563 blocked for more than 241 seconds.
[10392.326039] Tainted: P IOE 5.15.0-48-generic #54-Ubuntu
[10392.326042] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[10392.326045] task:rsync state:D stack: 0 pid:847563 ppid: 46824 flags:0x00004002
[10392.326051] Call Trace:
[10392.326054] <TASK>
[10392.326059] __schedule+0x23d/0x5a0
[10392.326070] schedule+0x4e/0xc0
[10392.326074] spl_panic+0xe7/0xe9 [spl]
[10392.326099] ? __cond_resched+0x1a/0x50
[10392.326103] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10392.326276] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10392.326425] ? dmu_buf_replace_user+0x65/0x80 [zfs]
[10392.326578] ? dmu_buf_set_user+0x13/0x20 [zfs]
[10392.326729] ? dmu_buf_set_user_ie+0x15/0x20 [zfs]
[10392.326882] zfs_znode_sa_init+0xdd/0xf0 [zfs]
[10392.327111] zfs_znode_alloc+0x101/0x5d0 [zfs]
[10392.327340] ? arc_space_consume+0x54/0x100 [zfs]
[10392.327479] ? dbuf_read_bonus+0x102/0x110 [zfs]
[10392.327630] ? RW_WRITE_HELD+0x9/0x30 [zfs]
[10392.327779] ? dmu_buf_unlock_parent+0x65/0xa0 [zfs]
[10392.327930] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10392.328080] ? dbuf_read_impl.constprop.0+0x352/0x490 [zfs]
[10392.328230] ? dbuf_rele_and_unlock+0x134/0x540 [zfs]
[10392.328379] ? __cond_resched+0x1a/0x50
[10392.328384] ? __cond_resched+0x1a/0x50
[10392.328387] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10392.328525] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10392.328662] ? __cond_resched+0x1a/0x50
[10392.328666] ? down_read+0x13/0xa0
[10392.328670] ? wmsum_add+0xe/0x20 [zfs]
[10392.328820] ? dbuf_read+0x1e6/0x5b0 [zfs]
[10392.328970] ? __cond_resched+0x1a/0x50
[10392.328974] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10392.329142] ? __raw_spin_unlock+0x9/0x10 [zfs]
[10392.329310] ? dnode_rele_and_unlock+0x69/0xf0 [zfs]
[10392.329479] ? do_raw_spin_unlock+0x9/0x10 [zfs]
[10392.329635] ? RW_WRITE_HELD+0x9/0x30 [zfs]
[10392.329819] ? dmu_object_info_from_dnode+0x92/0xa0 [zfs]
[10392.330064] zfs_zget+0x239/0x280 [zfs]
[10392.330289] zfs_dirent_lock+0x424/0x570 [zfs]
[10392.330519] zfs_dirlook+0x91/0x2d0 [zfs]
[10392.330748] zfs_lookup+0x25e/0x410 [zfs]
[10392.330977] zpl_lookup+0xcf/0x230 [zfs]
[10392.331206] ? capable_wrt_inode_uidgid+0x5f/0x70
[10392.331213] __lookup_slow+0x88/0x150
[10392.331219] walk_component+0x145/0x1c0
[10392.331223] ? zfs_dirty_inode+0x2af/0x330 [zfs]
[10392.331451] path_lookupat+0x6e/0x1c0
[10392.331456] filename_lookup+0xcf/0x1d0
[10392.331461] ? __check_object_size+0x1d/0x30
[10392.331466] ? strncpy_from_user+0x44/0x150
[10392.331473] ? getname_flags.part.0+0x4c/0x1b0
[10392.331478] user_path_at_empty+0x3f/0x60
[10392.331482] vfs_statx+0x7a/0x130
[10392.331488] __do_sys_newfstatat+0x36/0x70
[10392.331493] ? exit_to_user_mode_prepare+0x96/0xb0
[10392.331499] ? syscall_exit_to_user_mode+0x27/0x50
[10392.331505] ? do_syscall_64+0x69/0xc0
[10392.331509] ? exit_to_user_mode_prepare+0x37/0xb0
[10392.331513] ? syscall_exit_to_user_mode+0x27/0x50
[10392.331518] ? __x64_sys_newfstatat+0x1c/0x30
[10392.331522] ? do_syscall_64+0x69/0xc0
[10392.331525] __x64_sys_newfstatat+0x1c/0x30
[10392.331530] do_syscall_64+0x5c/0xc0
[10392.331533] ? do_syscall_64+0x69/0xc0
[10392.331536] ? syscall_exit_to_user_mode+0x27/0x50
[10392.331540] ? __x64_sys_close+0x11/0x50
[10392.331545] ? do_syscall_64+0x69/0xc0
[10392.331549] entry_SYSCALL_64_after_hwframe+0x61/0xcb
[10392.331555] RIP: 0033:0x7fbbb12afeee
[10392.331559] RSP: 002b:00007ffc2be9a998 EFLAGS: 00000246 ORIG_RAX: 0000000000000106
[10392.331563] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fbbb12afeee
[10392.331566] RDX: 00007ffc2be9aaa0 RSI: 00007ffc2be9bb30 RDI: 00000000ffffff9c
[10392.331569] RBP: 00007ffc2be9aaa0 R08: 00007ffc2be9bb2f R09: 0000000000000000
[10392.331571] R10: 0000000000000100 R11: 0000000000000246 R12: 0000000000000000
[10392.331574] R13: 00007ffc2be9bb30 R14: 000055776e8adfb0 R15: 00007ffc2be9aaa0
Sending dataset's snapshot as raw via zfs send -w pool1/dataset1@--2022-13-32--25-62-62--snapshot1 | zfs recv pool2/dataset2
works well, without errors.
Sending snapshots 2 & 3 zfs send -w pool1/dataset1@--2022-14-33--26-63-63--snapshot2 | zfs recv pool2/dataset3
gives errors in zfs and spl modules, but send/recv ends successfully.
We need to find out what file on your pool is corrupted and restore it from backup (actually, the data is fine, it is just metadata that is corrupt). A raw send would not check the contents, so the corruption just goes from one pool to another that way.
I am busy this week. Could you contact me on Monday at around 1pm in the OpenZFS slack channel? I could try to help you then.
I'm running into a panic in a similar situation. I suspect it may be the same cause
Arch linux (kernel Linux machine-name 6.0.6-arch1-1 #1 SMP PREEMPT_DYNAMIC Sat, 29 Oct 2022 14:08:39 +0000 x86_64 GNU/Linux)
2.1.6 (installed from non-official repo)
Kernel panic, unkillable zfs process during send/receive
Similar to the original report, but I'm doing an incremental send
zpool import pool1
zpool import pool2
zfs load-key pool1/dataset1
zfs send -I 'pool1/dataset1'@'2022-10-03' 'pool1/dataset1'@'2022-11-02' | zfs receive -s -F 'pool2/dataset1'
Dataset is encrypted. The panic happens if I only load the encryption key on the source and not the destination. It does not panic if the key is loaded on both ends
dmesg output
[ 1107.811733] VERIFY3(0 == zap_add(mos, dsl_dir_phys(pds)->dd_child_dir_zapobj, name, sizeof (uint64_t), 1, &ddobj, tx)) failed (0 == 17)
[ 1107.811740] PANIC at dsl_dir.c:951:dsl_dir_create_sync()
[ 1107.811742] Showing stack for process 2165
[ 1107.811743] CPU: 0 PID: 2165 Comm: txg_sync Tainted: P OE 6.0.6-arch1-1 #1 a46cc4b882cfc11c3bbb09d6a0fab3dcad53b5c2
[ 1107.811747] Hardware name: Dell Inc. Inspiron 7520/0PXH02, BIOS A11 02/20/2014
[ 1107.811748] Call Trace:
[ 1107.811752] <TASK>
[ 1107.811755] dump_stack_lvl+0x48/0x60
[ 1107.811762] spl_panic+0xf4/0x10c [spl 4584b05d58613f22c91dbd8e0bc55d01674607e6]
[ 1107.811776] dsl_dir_create_sync+0x25b/0x350 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.811885] dsl_dataset_create_sync+0x5d/0x3a0 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.811999] dmu_recv_begin_sync+0x390/0x9f0 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812070] dsl_sync_task_sync+0xac/0xf0 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812151] dsl_pool_sync+0x3f9/0x510 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812230] spa_sync+0x569/0xf90 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812325] ? spa_txg_history_init_io+0x117/0x120 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812451] txg_sync_thread+0x22b/0x3f0 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812576] ? txg_fini+0x270/0x270 [zfs dd8976dc3cdb4b6bb09fa0a98dc75c1b67476003]
[ 1107.812700] ? __thread_exit+0x20/0x20 [spl 4584b05d58613f22c91dbd8e0bc55d01674607e6]
[ 1107.812712] thread_generic_wrapper+0x5e/0x70 [spl 4584b05d58613f22c91dbd8e0bc55d01674607e6]
[ 1107.812725] kthread+0xde/0x110
[ 1107.812730] ? kthread_complete_and_exit+0x20/0x20
[ 1107.812734] ret_from_fork+0x22/0x30
[ 1107.812740] </TASK>
@i8-pi I think your issue isn't related to this one but a duplicate of #13598 and #14055.
I think you're correct. Sorry for the noise
Hello, I've permanent panic during
zfs send pool1/dataset1 | zfs recv -v -u -s pool2/dataset2
. It appears just after 179 GB of 183 GB is transfered. It happens on Ubuntu, Fedora, FreeBSD 13.1, OpenIndiana Hipster (OI-hipster-gui-20211031) and OmniOS (omnios-r151042).System information
Distribution Name = Ubuntu, Fedora, FreeBSD 13.1, OpenIndiana Hipster (OI-hipster-gui-20211031) and OmniOS (omnios-r151042) Distribution Version = Ubuntu 22.04, Fedora 36, FreeBSD 13.1, OpenIndiana Hipster (OI-hipster-gui-20211031) and OmniOS (omnios-r151042) Kernel Version = Linux hpws 5.15.0-48-generic #54-Ubuntu SMP Fri Aug 26 13:26:29 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux (as for Ubuntu 22.04) Architecture x86_64 OpenZFS Version= zfs-2.1.4-0ubuntu0.1, zfs-kmod-2.1.4-0ubuntu0.1 (as for Ubuntu 22.04)
Describe the problem you're observing:
I've permanent panic during
zfs send pool1/dataset1 | zfs recv -v -u -s pool2/dataset2
. It appears just after 179 GB of 183 GB is transfered. It also occurs not depending on if-v
and/or-s
are present or not (zfs send pool1/dataset1 | zfs recv -u pool2/dataset2
, and so on...)Source dataset is encrypted with ZFS native encryption (aes 256 gcm). Compression = lz4 Dedup = off Checksum = sha512
Target dataset: Compression = lz4 Dedup = off Checksum = on
Target dataset can be not encrypted or be on top of LUKS. Also tried to send via SSH, the panic occurs on sender's side. Despite the error occurs on the sender size,
zfs send pool1/dataset1 > /home/user/file.zfs
works without any error.UPD 1:
Just tried to
cat /home/user/file.zfs | zfs recv -v -u -s pool2/dataset2
, also got panic.UPD 2:
Also have errors in
rsync
and same panic trying to:zpool scrub pool1
andzpool scrub pool2
finished successfully without any error.UPD 3:
Sending dataset's snapshot as unencrypted raw via
zfs send -w pool1/dataset1@--2022-13-32--25-62-62--snapshot1 | zfs recv pool2/dataset2
works well, without errors. Sending snapshots 2 & 3zfs send -w pool1/dataset1@--2022-14-33--26-63-63--snapshot2 | zfs recv pool2/dataset3
gives errors in zfs and spl modules, but send/recv ends successfully.Describe how to reproduce the problem:
Include any warning/errors/backtraces from the system logs
Panic during
rsync
:gives errors: