Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

deadlock on multithreaded I/O #521

Closed
akorn opened this issue Jan 8, 2012 · 6 comments
Closed

deadlock on multithreaded I/O #521

akorn opened this issue Jan 8, 2012 · 6 comments
Milestone

Comments

@akorn
Copy link
Contributor

akorn commented Jan 8, 2012

Hi,

I've experienced a deadlock similar to #496; however, the stack traces are sufficiently different for me to think it's not the same issue.

My pool configuration is:

pool: tank
state: ONLINE
scan: scrub canceled on Sun Jan 8 01:21:49 2012
config:

    NAME         STATE     READ WRITE CKSUM
    tank         ONLINE       0     0     0
      raidz2-0   ONLINE       0     0     0
        sas1     ONLINE       0     0     0
        sas2     ONLINE       0     0     0
        sas3     ONLINE       0     0     0
        sas4     ONLINE       0     0     0
        sas5     ONLINE       0     0     0
    logs
      zfs_zil    ONLINE       0     0     0
    cache
      zfs_l2arc  ONLINE       0     0     0

errors: No known data errors

At the time of the deadlock, I was rsyncing (at a pathologically low speed) from a remote box, and several local processes were also accessing the pool. I/O load was not particularly high.

After the deadlock, it was impossible to log in (SSH would connect, but the shell would never start).

I'm running zfs 0.6.0.43-0ubuntu1~oneiric1 with a576bc0 merged and ZFS_LINK_MAX raised to ((0x1 << 31) - 1U) as suggested in #514.

Kernel log:

Jan  8 07:18:01 jayhawk kernel: INFO: rcu_sched_state detected stall on CPU 0 (t=6000 jiffies)
Jan  8 07:18:01 jayhawk kernel: sending NMI to all CPUs:
Jan  8 07:18:01 jayhawk kernel: NMI backtrace for cpu 0
Jan  8 07:18:01 jayhawk kernel: CPU 0
Jan  8 07:18:01 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 07:18:01 jayhawk kernel:
Jan  8 07:18:01 jayhawk kernel: Pid: 31058, comm: zfs_iput_taskq/ Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 07:18:01 jayhawk kernel: RIP: 0010:[<ffffffff8130cade>]  [<ffffffff8130cade>] find_next_bit+0x1e/0xc0
Jan  8 07:18:01 jayhawk kernel: RSP: 0018:ffff88041fc03d68  EFLAGS: 00000046
Jan  8 07:18:01 jayhawk kernel: RAX: 000000000000000c RBX: 0000000000000002 RCX: 0000000000000000
Jan  8 07:18:01 jayhawk kernel: RDX: 0000000000000002 RSI: 000000000000000c RDI: ffffffff81aaae80
Jan  8 07:18:01 jayhawk kernel: RBP: ffff88041fc03d68 R08: ffffffff81aaae80 R09: 0000000000000000
Jan  8 07:18:01 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: 000000000000d442
Jan  8 07:18:03 jayhawk kernel: R13: ffffffff81aaae80 R14: 0000000000000400 R15: 0000000000000096
Jan  8 07:18:03 jayhawk kernel: FS:  00007fdae3d17700(0000) GS:ffff88041fc00000(0000) knlGS:000000005611b960
Jan  8 07:18:03 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 07:18:03 jayhawk kernel: CR2: 0000000055f8f394 CR3: 00000003f93a8000 CR4: 00000000000006f0
Jan  8 07:18:03 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 07:18:03 jayhawk kernel: Process zfs_iput_taskq/ (pid: 31058, threadinfo ffff8803d5876000, task ffff8802fcc7c470)
Jan  8 07:18:03 jayhawk kernel: Stack:
Jan  8 07:18:03 jayhawk kernel:  ffff88041fc03db8 ffffffff81067dc2 ffff880400000001 0000000000000001
Jan  8 07:18:03 jayhawk kernel:  ffff88041fc03dd8 0000000000002710 ffffffff81a29280 ffffffff81a29280
Jan  8 07:18:03 jayhawk kernel:  ffff88041fc0df40 7fffffffffffffff ffff88041fc03dc8 ffffffff8106bd52
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  <IRQ>
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81067dc2>] default_send_IPI_mask_sequence_phys+0x52/0xe0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8106bd52>] physflat_send_IPI_all+0x12/0x20
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81068045>] arch_trigger_all_cpu_backtrace+0x55/0x80
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810fb856>] __rcu_pending+0x1b6/0x380
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810fbd75>] rcu_check_callbacks+0x65/0x120
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810a4b63>] update_process_times+0x43/0x80
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810d16ff>] tick_sched_timer+0x5f/0xb0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810b9853>] __run_hrtimer+0x63/0x1b0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810d16a0>] ? tick_nohz_handler+0xe0/0xe0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81053089>] ? read_tsc+0x9/0x20
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810ba083>] hrtimer_interrupt+0xe3/0x1f0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e4e2c>] ? call_softirq+0x1c/0x30
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81066f64>] smp_apic_timer_interrupt+0x64/0xa0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e409e>] apic_timer_interrupt+0x6e/0x80
Jan  8 07:18:03 jayhawk kernel:  <EOI>
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037aeea>] ? dbuf_rele_and_unlock+0x1a/0x1f0 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037b455>] dmu_buf_rele+0x25/0x30 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037bd4f>] dbuf_prefetch+0x1cf/0x280 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038e331>] ? dmu_zfetch+0x291/0xd40 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038ddc8>] dmu_zfetch_dofetch.isra.4+0xe8/0x160 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038e749>] dmu_zfetch+0x6a9/0xd40 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037ad71>] dbuf_read+0x5e1/0x740 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038fd5b>] dnode_hold_impl+0x13b/0x550 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa0390184>] dnode_hold+0x14/0x20 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa0380ff3>] dmu_object_info+0x23/0x60 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa03dd0dd>] zfs_unlinked_drain+0x7d/0x120 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e082c>] ? __schedule+0x3bc/0x890
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00edfd2>] taskq_thread+0x1e2/0x410 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00eddf0>] ? task_alloc+0x160/0x160 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810b5667>] kthread+0x87/0x90
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
Jan  8 07:18:03 jayhawk kernel: Code: e2 06 48 0f bd f6 48 8d 04 32 5d c3 90 55 48 39 f2 48 89 f0 48 89 e5 73 79 48 89 d1 49 89 d1 48 c1 e9 06 49 83 e1 c0 4c 8d 04 cf 
Jan  8 07:18:03 jayhawk kernel: <48> 89 f7 4c 29 cf 83 e2 3f 75 5f 48 f7 c7 c0 ff ff ff 74 2b 49 
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  <IRQ>  [<ffffffff81067dc2>] default_send_IPI_mask_sequence_phys+0x52/0xe0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8106bd52>] physflat_send_IPI_all+0x12/0x20
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81068045>] arch_trigger_all_cpu_backtrace+0x55/0x80
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810fb856>] __rcu_pending+0x1b6/0x380
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810fbd75>] rcu_check_callbacks+0x65/0x120
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810a4b63>] update_process_times+0x43/0x80
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810d16ff>] tick_sched_timer+0x5f/0xb0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810b9853>] __run_hrtimer+0x63/0x1b0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810d16a0>] ? tick_nohz_handler+0xe0/0xe0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81053089>] ? read_tsc+0x9/0x20
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810ba083>] hrtimer_interrupt+0xe3/0x1f0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e4e2c>] ? call_softirq+0x1c/0x30
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81066f64>] smp_apic_timer_interrupt+0x64/0xa0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e409e>] apic_timer_interrupt+0x6e/0x80
Jan  8 07:18:03 jayhawk kernel:  <EOI>  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037aeea>] ? dbuf_rele_and_unlock+0x1a/0x1f0 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037b455>] dmu_buf_rele+0x25/0x30 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037bd4f>] dbuf_prefetch+0x1cf/0x280 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038e331>] ? dmu_zfetch+0x291/0xd40 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038ddc8>] dmu_zfetch_dofetch.isra.4+0xe8/0x160 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038e749>] dmu_zfetch+0x6a9/0xd40 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa037ad71>] dbuf_read+0x5e1/0x740 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa038fd5b>] dnode_hold_impl+0x13b/0x550 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa0390184>] dnode_hold+0x14/0x20 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa0380ff3>] dmu_object_info+0x23/0x60 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa03dd0dd>] zfs_unlinked_drain+0x7d/0x120 [zfs]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e082c>] ? __schedule+0x3bc/0x890
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00edfd2>] taskq_thread+0x1e2/0x410 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
Jan  8 07:18:03 jayhawk kernel:  [<ffffffffa00eddf0>] ? task_alloc+0x160/0x160 [spl]
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810b5667>] kthread+0x87/0x90
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
Jan  8 07:18:03 jayhawk kernel: NMI backtrace for cpu 1
Jan  8 07:18:03 jayhawk kernel: CPU 1 
Jan  8 07:18:03 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 07:18:03 jayhawk kernel: 
Jan  8 07:18:03 jayhawk kernel: Pid: 0, comm: kworker/0:0 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 07:18:03 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 07:18:03 jayhawk kernel: RSP: 0000:ffff8804054cbeb8  EFLAGS: 00000246
Jan  8 07:18:03 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054cbfd8 RCX: 00000000ffffffff
Jan  8 07:18:03 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 07:18:03 jayhawk kernel: RBP: ffff8804054cbeb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel: R13: 0000000000000001 R14: 0000000000000000 R15: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: FS:  00002b78693086f0(0000) GS:ffff88041fc20000(0000) knlGS:000000005611b960
Jan  8 07:18:03 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 07:18:03 jayhawk kernel: CR2: 0000000055bf2000 CR3: 00000001eaa8c000 CR4: 00000000000006e0
Jan  8 07:18:03 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 07:18:03 jayhawk kernel: Process kworker/0:0 (pid: 0, threadinfo ffff8804054ca000, task ffff8804054d0000)
Jan  8 07:18:03 jayhawk kernel: Stack:
Jan  8 07:18:03 jayhawk kernel:  ffff8804054cbee8 ffffffff81054410 ffff8804054cbfd8 ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054cbf18 ffffffff8104aad6
Jan  8 07:18:03 jayhawk kernel:  ffff8804054cbf08 03f4fe87f591794a 0000000000000001 0000000000000000
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: NMI backtrace for cpu 2
Jan  8 07:18:03 jayhawk kernel: CPU 2 
Jan  8 07:18:03 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 07:18:03 jayhawk kernel: 
Jan  8 07:18:03 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 07:18:03 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 07:18:03 jayhawk kernel: RSP: 0018:ffff8804054e3eb8  EFLAGS: 00000246
Jan  8 07:18:03 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054e3fd8 RCX: 00000000ffffffff
Jan  8 07:18:03 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 07:18:03 jayhawk kernel: RBP: ffff8804054e3eb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel: R13: 0000000000000002 R14: 0000000000000000 R15: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: FS:  00007f10feb20700(0000) GS:ffff88041fc40000(0000) knlGS:000000005611b960
Jan  8 07:18:03 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 07:18:03 jayhawk kernel: CR2: 000000000805eb50 CR3: 0000000400c60000 CR4: 00000000000006e0
Jan  8 07:18:03 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 07:18:03 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff8804054e2000, task ffff8804054d5b40)
Jan  8 07:18:03 jayhawk kernel: Stack:
Jan  8 07:18:03 jayhawk kernel:  ffff8804054e3ee8 ffffffff81054410 ffff8804054e3fd8 ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054e3f18 ffffffff8104aad6
Jan  8 07:18:03 jayhawk kernel:  ffff8804054e3f08 0cf32962194b79f0 0000000000000002 0000000000000000
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: NMI backtrace for cpu 3
Jan  8 07:18:03 jayhawk kernel: CPU 3 
Jan  8 07:18:03 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 07:18:03 jayhawk kernel: 
Jan  8 07:18:03 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 07:18:03 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 07:18:03 jayhawk kernel: RSP: 0000:ffff8804054ffeb8  EFLAGS: 00000246
Jan  8 07:18:03 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054fffd8 RCX: 00000000ffffffff
Jan  8 07:18:03 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 07:18:03 jayhawk kernel: RBP: ffff8804054ffeb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel: R13: 0000000000000003 R14: 0000000000000000 R15: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: FS:  00007f266407c700(0000) GS:ffff88041fc60000(0000) knlGS:000000005611b960
Jan  8 07:18:03 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 07:18:03 jayhawk kernel: CR2: 00000000f7664a09 CR3: 0000000131795000 CR4: 00000000000006e0
Jan  8 07:18:03 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 07:18:03 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff8804054fe000, task ffff8804054ec470)
Jan  8 07:18:03 jayhawk kernel: Stack:
Jan  8 07:18:03 jayhawk kernel:  ffff8804054ffee8 ffffffff81054410 ffff8804054fffd8 ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054fff18 ffffffff8104aad6
Jan  8 07:18:03 jayhawk kernel:  ffff8804054fff08 b7b1c45da28780f9 0000000000000003 0000000000000000
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: NMI backtrace for cpu 4
Jan  8 07:18:03 jayhawk kernel: CPU 4 
Jan  8 07:18:03 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 07:18:03 jayhawk kernel: 
Jan  8 07:18:03 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 07:18:03 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 07:18:03 jayhawk kernel: RSP: 0018:ffff880405513eb8  EFLAGS: 00000246
Jan  8 07:18:03 jayhawk kernel: RAX: 0000000000000000 RBX: ffff880405513fd8 RCX: 00000000ffffffff
Jan  8 07:18:03 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 07:18:03 jayhawk kernel: RBP: ffff880405513eb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel: R13: 0000000000000004 R14: 0000000000000000 R15: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: FS:  00007f5cb68a67a0(0000) GS:ffff88041fc80000(0000) knlGS:0000000055849e70
Jan  8 07:18:03 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 07:18:03 jayhawk kernel: CR2: 00007fff0f8dedb8 CR3: 000000038454b000 CR4: 00000000000006e0
Jan  8 07:18:03 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 07:18:03 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff880405512000, task ffff88040550ada0)
Jan  8 07:18:03 jayhawk kernel: Stack:
Jan  8 07:18:03 jayhawk kernel:  ffff880405513ee8 ffffffff81054410 ffff880405513fd8 ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel:  0000000000000000 0000000000000000 ffff880405513f18 ffffffff8104aad6
Jan  8 07:18:03 jayhawk kernel:  ffff880405513f08 2ebceaa58ffe4b79 0000000000000004 0000000000000000
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: NMI backtrace for cpu 5
Jan  8 07:18:03 jayhawk kernel: CPU 5 
Jan  8 07:18:03 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 07:18:03 jayhawk kernel: 
Jan  8 07:18:03 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 07:18:03 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 07:18:03 jayhawk kernel: RSP: 0000:ffff88040552beb8  EFLAGS: 00000246
Jan  8 07:18:03 jayhawk kernel: RAX: 0000000000000000 RBX: ffff88040552bfd8 RCX: 00000000ffffffff
Jan  8 07:18:03 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 07:18:03 jayhawk kernel: RBP: ffff88040552beb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel: R13: 0000000000000005 R14: 0000000000000000 R15: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: FS:  00002aeec5586790(0000) GS:ffff88041fca0000(0000) knlGS:000000005b7fab70
Jan  8 07:18:03 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 07:18:03 jayhawk kernel: CR2: 00000000f75c9a09 CR3: 0000000316c7a000 CR4: 00000000000006e0
Jan  8 07:18:03 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 07:18:03 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 07:18:03 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff88040552a000, task ffff8804055216d0)
Jan  8 07:18:03 jayhawk kernel: Stack:
Jan  8 07:18:03 jayhawk kernel:  ffff88040552bee8 ffffffff81054410 ffff88040552bfd8 ffffffff81aaae80
Jan  8 07:18:03 jayhawk kernel:  0000000000000000 0000000000000000 ffff88040552bf18 ffffffff8104aad6
Jan  8 07:18:03 jayhawk kernel:  ffff88040552bf08 5e3fcc1570a6a5d4 0000000000000005 0000000000000000
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 07:18:03 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 07:18:03 jayhawk kernel: Call Trace:
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 07:18:03 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe

I don't know if the box actually recovered, but a similar message was logged a few hours later:

Jan  8 09:03:56 jayhawk kernel: INFO: rcu_sched_state detected stall on CPU 0 (t=6000 jiffies)
Jan  8 09:03:56 jayhawk kernel: sending NMI to all CPUs:
Jan  8 09:03:56 jayhawk kernel: NMI backtrace for cpu 0
Jan  8 09:03:56 jayhawk kernel: INFO: rcu_sched_state detected stalls on CPUs/tasks: { 0} (detected by 3, t=6002 jiffies)
Jan  8 09:03:56 jayhawk kernel: CPU 0 
Jan  8 09:03:56 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:03:56 jayhawk kernel: 
Jan  8 09:03:56 jayhawk kernel: Pid: 31058, comm: zfs_iput_taskq/ Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:03:56 jayhawk kernel: RIP: 0010:[<ffffffff8130cade>]  [<ffffffff8130cade>] find_next_bit+0x1e/0xc0
Jan  8 09:03:56 jayhawk kernel: RSP: 0018:ffff88041fc03d68  EFLAGS: 00000046
Jan  8 09:03:56 jayhawk kernel: RAX: 000000000000000c RBX: 0000000000000002 RCX: 0000000000000000
Jan  8 09:03:56 jayhawk kernel: RDX: 0000000000000002 RSI: 000000000000000c RDI: ffffffff81aaae80
Jan  8 09:03:56 jayhawk kernel: RBP: ffff88041fc03d68 R08: ffffffff81aaae80 R09: 0000000000000000
Jan  8 09:03:56 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: 000000000000d442
Jan  8 09:03:56 jayhawk kernel: R13: ffffffff81aaae80 R14: 0000000000000400 R15: 0000000000000096
Jan  8 09:03:56 jayhawk kernel: FS:  00007fdae3d17700(0000) GS:ffff88041fc00000(0000) knlGS:00000000556fa6c0
Jan  8 09:03:56 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:03:56 jayhawk kernel: CR2: 00000000ffee4e10 CR3: 0000000001a05000 CR4: 00000000000006f0
Jan  8 09:03:56 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:03:56 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:03:56 jayhawk kernel: Process zfs_iput_taskq/ (pid: 31058, threadinfo ffff8803d5876000, task ffff8802fcc7c470)
Jan  8 09:03:56 jayhawk kernel: Stack:
Jan  8 09:03:56 jayhawk kernel:  ffff88041fc03db8 ffffffff81067dc2 ffff880400000001 0000000000000001
Jan  8 09:03:56 jayhawk kernel:  ffff88041fc03dd8 0000000000002710 ffffffff81a29280 ffffffff81a29280
Jan  8 09:03:56 jayhawk kernel:  ffff88041fc0df40 7fffffffffffffff ffff88041fc03dc8 ffffffff8106bd52
Jan  8 09:03:56 jayhawk kernel: Call Trace:
Jan  8 09:03:56 jayhawk kernel:  <IRQ> 
Jan  8 09:03:56 jayhawk kernel:  [<ffffffff81067dc2>] default_send_IPI_mask_sequence_phys+0x52/0xe0
Jan  8 09:03:56 jayhawk kernel:  [<ffffffff8106bd52>] physflat_send_IPI_all+0x12/0x20
Jan  8 09:03:56 jayhawk kernel:  [<ffffffff81068045>] arch_trigger_all_cpu_backtrace+0x55/0x80
Jan  8 09:03:56 jayhawk kernel:  [<ffffffff810fb856>] __rcu_pending+0x1b6/0x380
Jan  8 09:03:56 jayhawk kernel:  [<ffffffff810fbd75>] rcu_check_callbacks+0x65/0x120
Jan  8 09:03:56 jayhawk kernel:  [<ffffffff810a4b63>] update_process_times+0x43/0x80
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810d16ff>] tick_sched_timer+0x5f/0xb0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810b9853>] __run_hrtimer+0x63/0x1b0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810d16a0>] ? tick_nohz_handler+0xe0/0xe0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81053089>] ? read_tsc+0x9/0x20
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810ba083>] hrtimer_interrupt+0xe3/0x1f0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e4e2c>] ? call_softirq+0x1c/0x30
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81066f64>] smp_apic_timer_interrupt+0x64/0xa0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e409e>] apic_timer_interrupt+0x6e/0x80
Jan  8 09:03:57 jayhawk kernel:  <EOI> 
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00ec1a5>] ? kmem_alloc_debug+0x65/0x120 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa037bb22>] dbuf_hold_impl+0x52/0xb0 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa037bc62>] dbuf_prefetch+0xe2/0x280 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038e331>] ? dmu_zfetch+0x291/0xd40 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038ddc8>] dmu_zfetch_dofetch.isra.4+0xe8/0x160 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038e749>] dmu_zfetch+0x6a9/0xd40 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa037ad71>] dbuf_read+0x5e1/0x740 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038fd5b>] dnode_hold_impl+0x13b/0x550 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa0390184>] dnode_hold+0x14/0x20 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa0380ff3>] dmu_object_info+0x23/0x60 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa03dd0dd>] zfs_unlinked_drain+0x7d/0x120 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810824f4>] ? finish_task_switch+0x44/0xe0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e082c>] ? __schedule+0x3bc/0x890
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00edfd2>] taskq_thread+0x1e2/0x410 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00eddf0>] ? task_alloc+0x160/0x160 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810b5667>] kthread+0x87/0x90
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
Jan  8 09:03:57 jayhawk kernel: Code: e2 06 48 0f bd f6 48 8d 04 32 5d c3 90 55 48 39 f2 48 89 f0 48 89 e5 73 79 48 89 d1 49 89 d1 48 c1 e9 06 49 83 e1 c0 4c 8d 04 cf 
Jan  8 09:03:57 jayhawk kernel: <48> 89 f7 4c 29 cf 83 e2 3f 75 5f 48 f7 c7 c0 ff ff ff 74 2b 49 
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  <IRQ>  [<ffffffff81067dc2>] default_send_IPI_mask_sequence_phys+0x52/0xe0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8106bd52>] physflat_send_IPI_all+0x12/0x20
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81068045>] arch_trigger_all_cpu_backtrace+0x55/0x80
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810fb856>] __rcu_pending+0x1b6/0x380
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810fbd75>] rcu_check_callbacks+0x65/0x120
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810a4b63>] update_process_times+0x43/0x80
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810d16ff>] tick_sched_timer+0x5f/0xb0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810b9853>] __run_hrtimer+0x63/0x1b0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810d16a0>] ? tick_nohz_handler+0xe0/0xe0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81053089>] ? read_tsc+0x9/0x20
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810ba083>] hrtimer_interrupt+0xe3/0x1f0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e4e2c>] ? call_softirq+0x1c/0x30
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81066f64>] smp_apic_timer_interrupt+0x64/0xa0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e409e>] apic_timer_interrupt+0x6e/0x80
Jan  8 09:03:57 jayhawk kernel:  <EOI>  [<ffffffffa00ec1a5>] ? kmem_alloc_debug+0x65/0x120 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa037bb22>] dbuf_hold_impl+0x52/0xb0 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa037bc62>] dbuf_prefetch+0xe2/0x280 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038e331>] ? dmu_zfetch+0x291/0xd40 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038ddc8>] dmu_zfetch_dofetch.isra.4+0xe8/0x160 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038e749>] dmu_zfetch+0x6a9/0xd40 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa037ad71>] dbuf_read+0x5e1/0x740 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa038fd5b>] dnode_hold_impl+0x13b/0x550 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa0390184>] dnode_hold+0x14/0x20 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa0380ff3>] dmu_object_info+0x23/0x60 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa03dd0dd>] zfs_unlinked_drain+0x7d/0x120 [zfs]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810824f4>] ? finish_task_switch+0x44/0xe0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e082c>] ? __schedule+0x3bc/0x890
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00edfd2>] taskq_thread+0x1e2/0x410 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
Jan  8 09:03:57 jayhawk kernel:  [<ffffffffa00eddf0>] ? task_alloc+0x160/0x160 [spl]
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810b5667>] kthread+0x87/0x90
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
Jan  8 09:03:57 jayhawk kernel: NMI backtrace for cpu 1
Jan  8 09:03:57 jayhawk kernel: CPU 1 
Jan  8 09:03:57 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:03:57 jayhawk kernel: 
Jan  8 09:03:57 jayhawk kernel: Pid: 0, comm: kworker/0:0 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:03:57 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:03:57 jayhawk kernel: RSP: 0018:ffff8804054cbeb8  EFLAGS: 00000246
Jan  8 09:03:57 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054cbfd8 RCX: 00000000ffffffff
Jan  8 09:03:57 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 09:03:57 jayhawk kernel: RBP: ffff8804054cbeb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel: R13: 0000000000000001 R14: 0000000000000000 R15: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: FS:  00002b78693086f0(0000) GS:ffff88041fc20000(0000) knlGS:000000005611b960
Jan  8 09:03:57 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:03:57 jayhawk kernel: CR2: 0000000000619000 CR3: 000000011e2fd000 CR4: 00000000000006e0
Jan  8 09:03:57 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:03:57 jayhawk kernel: Process kworker/0:0 (pid: 0, threadinfo ffff8804054ca000, task ffff8804054d0000)
Jan  8 09:03:57 jayhawk kernel: Stack:
Jan  8 09:03:57 jayhawk kernel:  ffff8804054cbee8 ffffffff81054410 ffff8804054cbfd8 ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054cbf18 ffffffff8104aad6
Jan  8 09:03:57 jayhawk kernel:  ffff8804054cbf08 03f4fe87f591794a 0000000000000001 0000000000000000
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: NMI backtrace for cpu 2
Jan  8 09:03:57 jayhawk kernel: CPU 2 
Jan  8 09:03:57 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:03:57 jayhawk kernel: 
Jan  8 09:03:57 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:03:57 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:03:57 jayhawk kernel: RSP: 0018:ffff8804054e3eb8  EFLAGS: 00000246
Jan  8 09:03:57 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054e3fd8 RCX: 00000000ffffffff
Jan  8 09:03:57 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 09:03:57 jayhawk kernel: RBP: ffff8804054e3eb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel: R13: 0000000000000002 R14: 0000000000000000 R15: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: FS:  00002b653981a6f0(0000) GS:ffff88041fc40000(0000) knlGS:0000000055849e70
Jan  8 09:03:57 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:03:57 jayhawk kernel: CR2: 00002b6539566850 CR3: 000000038454b000 CR4: 00000000000006e0
Jan  8 09:03:57 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:03:57 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff8804054e2000, task ffff8804054d5b40)
Jan  8 09:03:57 jayhawk kernel: Stack:
Jan  8 09:03:57 jayhawk kernel:  ffff8804054e3ee8 ffffffff81054410 ffff8804054e3fd8 ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054e3f18 ffffffff8104aad6
Jan  8 09:03:57 jayhawk kernel:  ffff8804054e3f08 0cf32962194b79f0 0000000000000002 0000000000000000
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: NMI backtrace for cpu 3
Jan  8 09:03:57 jayhawk kernel: CPU 3 
Jan  8 09:03:57 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:03:57 jayhawk kernel: 
Jan  8 09:03:57 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:03:57 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:03:57 jayhawk kernel: RSP: 0000:ffff8804054ffeb8  EFLAGS: 00000246
Jan  8 09:03:57 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054fffd8 RCX: 00000000ffffffff
Jan  8 09:03:57 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 09:03:57 jayhawk kernel: RBP: ffff8804054ffeb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel: R13: 0000000000000003 R14: 0000000000000000 R15: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: FS:  00002b78693086f0(0000) GS:ffff88041fc60000(0000) knlGS:000000005611b960
Jan  8 09:03:57 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:03:57 jayhawk kernel: CR2: 00000000012bc38c CR3: 00000001eaa8c000 CR4: 00000000000006e0
Jan  8 09:03:57 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:03:57 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff8804054fe000, task ffff8804054ec470)
Jan  8 09:03:57 jayhawk kernel: Stack:
Jan  8 09:03:57 jayhawk kernel:  ffff8804054ffee8 ffffffff81054410 ffff8804054fffd8 ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054fff18 ffffffff8104aad6
Jan  8 09:03:57 jayhawk kernel:  ffff8804054fff08 b7b1c45da28780f9 0000000000000003 0000000000000000
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: NMI backtrace for cpu 4
Jan  8 09:03:57 jayhawk kernel: CPU 4 
Jan  8 09:03:57 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:03:57 jayhawk kernel: 
Jan  8 09:03:57 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:03:57 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:03:57 jayhawk kernel: RSP: 0000:ffff880405513eb8  EFLAGS: 00000246
Jan  8 09:03:57 jayhawk kernel: RAX: 0000000000000000 RBX: ffff880405513fd8 RCX: 00000000ffffffff
Jan  8 09:03:57 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 09:03:57 jayhawk kernel: RBP: ffff880405513eb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel: R13: 0000000000000004 R14: 0000000000000000 R15: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: FS:  00007f266407c700(0000) GS:ffff88041fc80000(0000) knlGS:00000000f75c98d0
Jan  8 09:03:57 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:03:57 jayhawk kernel: CR2: 00000000f7686c6a CR3: 0000000001a05000 CR4: 00000000000006e0
Jan  8 09:03:57 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:03:57 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff880405512000, task ffff88040550ada0)
Jan  8 09:03:57 jayhawk kernel: Stack:
Jan  8 09:03:57 jayhawk kernel:  ffff880405513ee8 ffffffff81054410 ffff880405513fd8 ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel:  0000000000000000 0000000000000000 ffff880405513f18 ffffffff8104aad6
Jan  8 09:03:57 jayhawk kernel:  ffff880405513f08 2ebceaa58ffe4b79 0000000000000004 0000000000000000
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: NMI backtrace for cpu 5
Jan  8 09:03:57 jayhawk kernel: CPU 5 
Jan  8 09:03:57 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:03:57 jayhawk kernel: 
Jan  8 09:03:57 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:03:57 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:03:57 jayhawk kernel: RSP: 0018:ffff88040552beb8  EFLAGS: 00000246
Jan  8 09:03:57 jayhawk kernel: RAX: 0000000000000000 RBX: ffff88040552bfd8 RCX: 00000000ffffffff
Jan  8 09:03:57 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 09:03:57 jayhawk kernel: RBP: ffff88040552beb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel: R13: 0000000000000005 R14: 0000000000000000 R15: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: FS:  00007f10feb20700(0000) GS:ffff88041fca0000(0000) knlGS:00000000f76516c0
Jan  8 09:03:57 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:03:57 jayhawk kernel: CR2: 0000000009033208 CR3: 0000000400c60000 CR4: 00000000000006e0
Jan  8 09:03:57 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:03:57 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:03:57 jayhawk kernel: Process kworker/0:1 (pid: 0, threadinfo ffff88040552a000, task ffff8804055216d0)
Jan  8 09:03:57 jayhawk kernel: Stack:
Jan  8 09:03:57 jayhawk kernel:  ffff88040552bee8 ffffffff81054410 ffff88040552bfd8 ffffffff81aaae80
Jan  8 09:03:57 jayhawk kernel:  0000000000000000 0000000000000000 ffff88040552bf18 ffffffff8104aad6
Jan  8 09:03:57 jayhawk kernel:  ffff88040552bf08 5e3fcc1570a6a5d4 0000000000000005 0000000000000000
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:03:57 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 09:03:57 jayhawk kernel: Call Trace:
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:03:57 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe

And then again:

Jan  8 09:11:54 jayhawk kernel: INFO: rcu_sched_state detected stall on CPU 0 (t=6000 jiffies)
Jan  8 09:11:54 jayhawk kernel: sending NMI to all CPUs:
Jan  8 09:11:54 jayhawk kernel: NMI backtrace for cpu 0
Jan  8 09:11:54 jayhawk kernel: CPU 0 
Jan  8 09:11:54 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:11:54 jayhawk kernel: 
Jan  8 09:11:54 jayhawk kernel: Pid: 31058, comm: zfs_iput_taskq/ Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:11:54 jayhawk kernel: RIP: 0010:[<ffffffff8130cade>]  [<ffffffff8130cade>] find_next_bit+0x1e/0xc0
Jan  8 09:11:54 jayhawk kernel: RSP: 0018:ffff88041fc03d68  EFLAGS: 00000046
Jan  8 09:11:54 jayhawk kernel: RAX: 000000000000000c RBX: 0000000000000002 RCX: 0000000000000000
Jan  8 09:11:54 jayhawk kernel: RDX: 0000000000000002 RSI: 000000000000000c RDI: ffffffff81aaae80
Jan  8 09:11:55 jayhawk kernel: RBP: ffff88041fc03d68 R08: ffffffff81aaae80 R09: 0000000000000000
Jan  8 09:11:55 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: 000000000000d442
Jan  8 09:11:55 jayhawk kernel: R13: ffffffff81aaae80 R14: 0000000000000400 R15: 0000000000000096
Jan  8 09:11:55 jayhawk kernel: FS:  00007ffe3b8257c0(0000) GS:ffff88041fc00000(0000) knlGS:00000000f76798d0
Jan  8 09:11:55 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:11:55 jayhawk kernel: CR2: 00002ae456fd48f0 CR3: 0000000001a05000 CR4: 00000000000006f0
Jan  8 09:11:55 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:11:55 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:11:55 jayhawk kernel: Process zfs_iput_taskq/ (pid: 31058, threadinfo ffff8803d5876000, task ffff8802fcc7c470)
Jan  8 09:11:55 jayhawk kernel: Stack:
Jan  8 09:11:55 jayhawk kernel:  ffff88041fc03db8 ffffffff81067dc2 ffff880400000001 0000000000000001
Jan  8 09:11:55 jayhawk kernel:  ffff88041fc03dd8 0000000000002710 ffffffff81a29280 ffffffff81a29280
Jan  8 09:11:55 jayhawk kernel:  ffff88041fc0df40 7fffffffffffffff ffff88041fc03dc8 ffffffff8106bd52
Jan  8 09:11:55 jayhawk kernel: Call Trace:
Jan  8 09:11:55 jayhawk kernel:  <IRQ> 
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81067dc2>] default_send_IPI_mask_sequence_phys+0x52/0xe0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8106bd52>] physflat_send_IPI_all+0x12/0x20
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81068045>] arch_trigger_all_cpu_backtrace+0x55/0x80
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810fb856>] __rcu_pending+0x1b6/0x380
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81167a66>] ? kfree+0x36/0x130
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810fbd75>] rcu_check_callbacks+0x65/0x120
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810a4b63>] update_process_times+0x43/0x80
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810d16ff>] tick_sched_timer+0x5f/0xb0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810b9853>] __run_hrtimer+0x63/0x1b0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8109b574>] ? __do_softirq+0xf4/0x200
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810d16a0>] ? tick_nohz_handler+0xe0/0xe0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81053089>] ? read_tsc+0x9/0x20
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810ba083>] hrtimer_interrupt+0xe3/0x1f0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81066f64>] smp_apic_timer_interrupt+0x64/0xa0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815e409e>] apic_timer_interrupt+0x6e/0x80
Jan  8 09:11:55 jayhawk kernel:  <EOI> 
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00ec06c>] ? kmem_free_debug+0xc/0x20 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa037bb63>] dbuf_hold_impl+0x93/0xb0 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa037bc62>] dbuf_prefetch+0xe2/0x280 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038e99c>] ? dmu_zfetch+0x8fc/0xd40 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038ddc8>] dmu_zfetch_dofetch.isra.4+0xe8/0x160 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038e749>] dmu_zfetch+0x6a9/0xd40 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa037ad71>] dbuf_read+0x5e1/0x740 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038fd5b>] dnode_hold_impl+0x13b/0x550 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa0390184>] dnode_hold+0x14/0x20 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa0380ff3>] dmu_object_info+0x23/0x60 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa03dd0dd>] zfs_unlinked_drain+0x7d/0x120 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00edfd2>] taskq_thread+0x1e2/0x410 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00eddf0>] ? task_alloc+0x160/0x160 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810b5667>] kthread+0x87/0x90
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
Jan  8 09:11:55 jayhawk kernel: Code: e2 06 48 0f bd f6 48 8d 04 32 5d c3 90 55 48 39 f2 48 89 f0 48 89 e5 73 79 48 89 d1 49 89 d1 48 c1 e9 06 49 83 e1 c0 4c 8d 04 cf 
Jan  8 09:11:55 jayhawk kernel: <48> 89 f7 4c 29 cf 83 e2 3f 75 5f 48 f7 c7 c0 ff ff ff 74 2b 49 
Jan  8 09:11:55 jayhawk kernel: Call Trace:
Jan  8 09:11:55 jayhawk kernel:  <IRQ>  [<ffffffff81067dc2>] default_send_IPI_mask_sequence_phys+0x52/0xe0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8106bd52>] physflat_send_IPI_all+0x12/0x20
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81068045>] arch_trigger_all_cpu_backtrace+0x55/0x80
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810fb856>] __rcu_pending+0x1b6/0x380
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81167a66>] ? kfree+0x36/0x130
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810fbd75>] rcu_check_callbacks+0x65/0x120
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810a4b63>] update_process_times+0x43/0x80
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810d16ff>] tick_sched_timer+0x5f/0xb0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810b9853>] __run_hrtimer+0x63/0x1b0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8109b574>] ? __do_softirq+0xf4/0x200
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810d16a0>] ? tick_nohz_handler+0xe0/0xe0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81053089>] ? read_tsc+0x9/0x20
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810ba083>] hrtimer_interrupt+0xe3/0x1f0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81066f64>] smp_apic_timer_interrupt+0x64/0xa0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815e409e>] apic_timer_interrupt+0x6e/0x80
Jan  8 09:11:55 jayhawk kernel:  <EOI>  [<ffffffffa00ec06c>] ? kmem_free_debug+0xc/0x20 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa037bb63>] dbuf_hold_impl+0x93/0xb0 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa037bc62>] dbuf_prefetch+0xe2/0x280 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038e99c>] ? dmu_zfetch+0x8fc/0xd40 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038ddc8>] dmu_zfetch_dofetch.isra.4+0xe8/0x160 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038e749>] dmu_zfetch+0x6a9/0xd40 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa037ad71>] dbuf_read+0x5e1/0x740 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00ec071>] ? kmem_free_debug+0x11/0x20 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa038fd5b>] dnode_hold_impl+0x13b/0x550 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa0390184>] dnode_hold+0x14/0x20 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa0380ff3>] dmu_object_info+0x23/0x60 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa03dd0dd>] zfs_unlinked_drain+0x7d/0x120 [zfs]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00edfd2>] taskq_thread+0x1e2/0x410 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
Jan  8 09:11:55 jayhawk kernel:  [<ffffffffa00eddf0>] ? task_alloc+0x160/0x160 [spl]
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810b5667>] kthread+0x87/0x90
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
Jan  8 09:11:55 jayhawk kernel: NMI backtrace for cpu 1
Jan  8 09:11:55 jayhawk kernel: CPU 1 
Jan  8 09:11:55 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:11:55 jayhawk kernel: 
Jan  8 09:11:55 jayhawk kernel: Pid: 0, comm: kworker/0:0 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:11:55 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:11:55 jayhawk kernel: RSP: 0000:ffff8804054cbeb8  EFLAGS: 00000246
Jan  8 09:11:55 jayhawk kernel: RAX: 0000000000000000 RBX: ffff8804054cbfd8 RCX: 00000000ffffffff
Jan  8 09:11:55 jayhawk kernel: RDX: 0000000000000000 RSI: 0000000000000001 RDI: ffffffff81b57fa8
Jan  8 09:11:55 jayhawk kernel: RBP: ffff8804054cbeb8 R08: 0000000000000000 R09: 0000000000000000
Jan  8 09:11:55 jayhawk kernel: R10: 0000000000000000 R11: 0000000000000000 R12: ffffffff81aaae80
Jan  8 09:11:55 jayhawk kernel: R13: 0000000000000001 R14: 0000000000000000 R15: 0000000000000000
Jan  8 09:11:55 jayhawk kernel: FS:  00002ad32c5ea6f0(0000) GS:ffff88041fc20000(0000) knlGS:000000005b8fab70
Jan  8 09:11:55 jayhawk kernel: CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
Jan  8 09:11:55 jayhawk kernel: CR2: 0000000055576000 CR3: 0000000316c7a000 CR4: 00000000000006e0
Jan  8 09:11:55 jayhawk kernel: DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
Jan  8 09:11:55 jayhawk kernel: DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Jan  8 09:11:55 jayhawk kernel: Process kworker/0:0 (pid: 0, threadinfo ffff8804054ca000, task ffff8804054d0000)
Jan  8 09:11:55 jayhawk kernel: Stack:
Jan  8 09:11:55 jayhawk kernel:  ffff8804054cbee8 ffffffff81054410 ffff8804054cbfd8 ffffffff81aaae80
Jan  8 09:11:55 jayhawk kernel:  0000000000000000 0000000000000000 ffff8804054cbf18 ffffffff8104aad6
Jan  8 09:11:55 jayhawk kernel:  ffff8804054cbf08 03f4fe87f591794a 0000000000000001 0000000000000000
Jan  8 09:11:55 jayhawk kernel: Call Trace:
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:11:55 jayhawk kernel: Code: 00 00 00 00 00 55 48 89 e5 fa 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb 5d c3 66 0f 1f 84 00 00 00 00 00 55 48 89 e5 fb f4 <5d> c3 0f 1f 84 00 00 00 00 00 55 48 89 e5 f4 5d c3 66 0f 1f 84 
Jan  8 09:11:55 jayhawk kernel: Call Trace:
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff81054410>] default_idle+0x40/0x100
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff8104aad6>] cpu_idle+0xc6/0x110
Jan  8 09:11:55 jayhawk kernel:  [<ffffffff815cf1f1>] start_secondary+0x1f7/0x1fe
Jan  8 09:11:55 jayhawk kernel: NMI backtrace for cpu 2
Jan  8 09:11:55 jayhawk kernel: INFO: rcu_sched_state detected stalls on CPUs/tasks: { 0} (detected by 1, t=6073 jiffies)
Jan  8 09:11:55 jayhawk kernel: CPU 2 
Jan  8 09:11:55 jayhawk kernel: Modules linked in: zfs(P) zcommon(P) znvpair(P) zavl(P) zunicode(P) spl loop ipt_LOG xt_limit xt_tcpudp xt_state iptable_nat nf_nat nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 iptable_filter ip_tables x_tables nfsd lockd nfs_acl auth_rpcgss sunrpc dummy fuse zlib_deflate sp5100_tco i2c_piix4 k10temp amd64_edac_mod hwmon edac_core i2c_core button ata_generic pata_acpi mpt2sas pata_atiixp scsi_transport_sas raid_class e1000e [last unloaded: spl]
Jan  8 09:11:55 jayhawk kernel: 
Jan  8 09:11:55 jayhawk kernel: Pid: 0, comm: kworker/0:1 Tainted: P            3.1.6-vs2.3.2.5-newjayhawk #4 Supermicro H8DCL/H8DCL
Jan  8 09:11:55 jayhawk kernel: RIP: 0010:[<ffffffff8106e8e6>]  [<ffffffff8106e8e6>] native_safe_halt+0x6/0x10
Jan  8 09:11:55 jayhawk kernel: RSP: 0000:ffff8804054e3eb8  EFL
@akorn
Copy link
Contributor Author

akorn commented Jan 8, 2012

OK, based on the timing, let me amend my statement about there not being heavy i/o.

I have 9 vservers (as in Linux-vserver) running on this box, and all of their cron.weekly jobs were executing simultaneously (and those of the host itself as well).

Their rootfs is on a zvol with ext4 on top; their other filesystems (like /var and /tmp) are zfs.

I'm now running the same jobs simultaneously on purpose, and am observing high load (50+) and very high i/o latency (it takes about a minute to start top(1)). The system is otherwise responsive.

It's unclear what's causing the load; if I have top hide idle tasks, there are almost no processes (certainly nowhere near 50) being displayed.

Not very surprisingly, there are extended periods (like 10-15 seconds) during which zpool iostat -v reports zero activity; iostat -x agrees.

As I watch, the periods with any i/o activity are becoming fewer and farther between until no further i/o appears to take place at all. The high load remains.

Some sample output:

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
           0.33    0.00   17.17   32.67    0.00   49.83

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util
sda               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdb               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdc               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdd               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdh               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdg               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdi               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sdf               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
sde               0.00     0.00    0.00    0.00     0.00     0.00     0.00     0.00    0.00    0.00    0.00   0.00   0.00
                capacity     operations    bandwidth
pool         alloc   free   read  write   read  write
-----------  -----  -----  -----  -----  -----  -----
tank         3.55T  5.52T      0      0      0      0
  raidz2     3.55T  5.52T      0      0      0      0
    sas1         -      -      0      0      0      0
    sas2         -      -      0      0      0      0
    sas3         -      -      0      0      0      0
    sas4         -      -      0      0      0      0
    sas5         -      -      0      0      0      0
  zfs_zil     688K  7.94G      0      0      0      0
cache            -      -      -      -      -      -
  zfs_l2arc  9.75G   110G      0      0      0      0
-----------  -----  -----  -----  -----  -----  -----

output from top(1), with threads on, idle processes filtered:

top - 18:27:57 up  6:30,  0 users,  load average: 72.32, 58.38, 38.58
Threads: 1124 total,  12 running, 1111 sleeping,   0 stopped,   1 zombie
%Cpu0  :  0.0 us, 75.4 sy,  0.0 ni, 24.6 id,  0.0 wa,  0.0 hi,  0.0 si,  0.0 st
%Cpu1  :  0.3 us,  2.3 sy,  0.3 ni,  0.0 id, 97.1 wa,  0.0 hi,  0.0 si,  0.0 st
%Cpu2  :  0.0 us,  1.0 sy,  0.0 ni, 99.0 id,  0.0 wa,  0.0 hi,  0.0 si,  0.0 st
%Cpu3  :  0.0 us,  0.2 sy,  0.0 ni, 24.6 id, 75.2 wa,  0.0 hi,  0.0 si,  0.0 st
%Cpu4  :  0.0 us,  0.0 sy,  0.0 ni,100.0 id,  0.0 wa,  0.0 hi,  0.0 si,  0.0 st
%Cpu5  :  0.0 us,  0.3 sy,  0.0 ni, 99.7 id,  0.0 wa,  0.0 hi,  0.0 si,  0.0 st
Mb Mem:     15986 total,     9132 used,     6853 free,      148 buffers
Mb Swap:     4095 total,        0 used,     4095 free,      398 cached

  PID USER      NI  VIRT  RES SWAP  SHR S nTH  %CPU %MEM    TIME+  WCHAN                                             COMMAND                                           
 9306 root       0 14168 2356    0  948 R   1   2.3  0.0   0:07.48 -                                                 top                                               
 4116 root       0     0    0    0    0 S   1   0.3  0.0   1:37.28 worker_thread                                     [kworker/1:1]                                     
 9292 build     10  2088  528    0  436 S   1   0.3  0.0   0:01.72 pipe_wait                                         awk -WInteractive -v events=1 -v wait=0  .($2 ~ " 
10291 root       0  9608  860    0  716 S   1   0.3  0.0   0:00.17 pause                                             iostat -x 3 /dev/sda /dev/sdb /dev/sdc /dev/sdd / 
10695 postgres   0 42264 2004    0 1252 S   1   0.3  0.0   0:04.63 poll_schedule_timeout                             postgres: writer process                          
26894 root       0     0    0    0    0 S   1   0.3  0.0   0:40.28 worker_thread                                     [kworker/2:2]                                     

I tried to get a process dump using echo t >/proc/sysrq-trigger, but netconsole somehow breaks it up to the point that it's almost useless.

Anyway, here goes what I could salvage (only processes in D state):

2012-01-08T18:28:59+01:00      D
2012-01-08T18:28:59+01:00  1973      2 0x00000000
2012-01-08T18:28:59+01:00 Trace:
2012-01-08T18:28:59+01:00 ? default_wake_function+0xd/0x10
2012-01-08T18:28:59+01:00 schedule+0x3a/0x50
2012-01-08T18:28:59+01:00 cv_wait_common+0x72/0xd0 [spl]
2012-01-08T18:28:59+01:00 ? add_wait_queue+0x60/0x60
2012-01-08T18:28:59+01:00 __cv_wait+0xe/0x10 [spl]
2012-01-08T18:28:59+01:00 zio_wait+0xe3/0x150 [zfs]
2012-01-08T18:28:59+01:00 l2arc_feed_thread+0x656/0x840 [zfs]
2012-01-08T18:28:59+01:00 ? l2arc_evict+0x300/0x300 [zfs]
2012-01-08T18:28:59+01:00 ? l2arc_evict+0x300/0x300 [zfs]
2012-01-08T18:28:59+01:00 thread_generic_wrapper+0x73/0x90 [spl]
2012-01-08T18:28:59+01:00 ? __thread_create+0x160/0x160 [spl]
2012-01-08T18:28:59+01:00 kthread+0x87/0x90
2012-01-08T18:28:59+01:00 kernel_thread_helper+0x4/0x10
2012-01-08T18:29:00+01:00 ? flush_kthread_worker+0xa0/0xa0
2012-01-08T18:29:00+01:00 ? gs_change+0x13/0x13

2012-01-08T18:28:59+01:00      D
2012-01-08T18:28:59+01:00  1973      2 0x00000000
2012-01-08T18:28:59+01:00 Trace:
2012-01-08T18:28:59+01:00 ? default_wake_function+0xd/0x10
2012-01-08T18:28:59+01:00 schedule+0x3a/0x50
2012-01-08T18:28:59+01:00 cv_wait_common+0x72/0xd0 [spl]
2012-01-08T18:28:59+01:00 ? add_wait_queue+0x60/0x60
2012-01-08T18:28:59+01:00 __cv_wait+0xe/0x10 [spl]
2012-01-08T18:28:59+01:00 zio_wait+0xe3/0x150 [zfs]
2012-01-08T18:28:59+01:00 l2arc_feed_thread+0x656/0x840 [zfs]
2012-01-08T18:28:59+01:00 ? l2arc_evict+0x300/0x300 [zfs]
2012-01-08T18:28:59+01:00 ? l2arc_evict+0x300/0x300 [zfs]
2012-01-08T18:28:59+01:00 thread_generic_wrapper+0x73/0x90 [spl]
2012-01-08T18:28:59+01:00 ? __thread_create+0x160/0x160 [spl]
2012-01-08T18:28:59+01:00 kthread+0x87/0x90
2012-01-08T18:28:59+01:00 kernel_thread_helper+0x4/0x10
2012-01-08T18:29:00+01:00 ? flush_kthread_worker+0xa0/0xa0
2012-01-08T18:29:00+01:00 ? gs_change+0x13/0x13

2012-01-08T18:29:00+01:00          D
2012-01-08T18:29:00+01:00  1978      2 0x00000000
2012-01-08T18:29:00+01:00 Trace:
2012-01-08T18:29:00+01:00 schedule+0x3a/0x50
2012-01-08T18:29:00+01:00 cv_wait_common+0x72/0xd0 [spl]
2012-01-08T18:29:00+01:00 ? add_wait_queue+0x60/0x60
2012-01-08T18:29:00+01:00 __cv_wait+0xe/0x10 [spl]
2012-01-08T18:29:00+01:00 txg_wait_open+0x73/0xa0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_wait+0xdd/0xe0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_assign+0x5e/0x410 [zfs]
2012-01-08T18:29:00+01:00 ? dsl_dataset_block_freeable+0x3b/0x50 [zfs]
2012-01-08T18:29:00+01:00 zvol_write+0x7c/0x380 [zfs]
2012-01-08T18:29:00+01:00 taskq_thread+0x1e2/0x410 [spl]
2012-01-08T18:29:00+01:00 ? try_to_wake_up+0x290/0x290
2012-01-08T18:29:00+01:00 ? task_alloc+0x160/0x160 [spl]
2012-01-08T18:29:00+01:00 kthread+0x87/0x90
2012-01-08T18:29:00+01:00 kernel_thread_helper+0x4/0x10
2012-01-08T18:29:00+01:00 ? flush_kthread_worker+0xa0/0xa0
2012-01-08T18:29:00+01:00 ? gs_change+0x13/0x13

2012-01-08T18:29:00+01:00          D
2012-01-08T18:29:00+01:00  1981      2 0x00000000
2012-01-08T18:29:00+01:00 Trace:
2012-01-08T18:29:00+01:00 schedule+0x3a/0x50
2012-01-08T18:29:00+01:00 cv_wait_common+0x72/0xd0 [spl]
2012-01-08T18:29:00+01:00 ? add_wait_queue+0x60/0x60
2012-01-08T18:29:00+01:00 __cv_wait+0xe/0x10 [spl]
2012-01-08T18:29:00+01:00 txg_wait_open+0x73/0xa0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_wait+0xdd/0xe0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_assign+0x5e/0x410 [zfs]
2012-01-08T18:29:00+01:00 ? dsl_dataset_block_freeable+0x3b/0x50 [zfs]
2012-01-08T18:29:00+01:00 zvol_write+0x7c/0x380 [zfs]
2012-01-08T18:29:00+01:00 taskq_thread+0x1e2/0x410 [spl]
2012-01-08T18:29:00+01:00 ? try_to_wake_up+0x290/0x290
2012-01-08T18:29:00+01:00 ? task_alloc+0x160/0x160 [spl]
2012-01-08T18:29:00+01:00 kthread+0x87/0x90
2012-01-08T18:29:00+01:00 kernel_thread_helper+0x4/0x10
2012-01-08T18:29:00+01:00 ? flush_kthread_worker+0xa0/0xa0
2012-01-08T18:29:00+01:00 ? gs_change+0x13/0x13
2012-01-08T18:29:00+01:00          D
2012-01-08T18:29:00+01:00  1982      2 0x00000000
2012-01-08T18:29:00+01:00 Trace:
2012-01-08T18:29:00+01:00 schedule+0x3a/0x50
2012-01-08T18:29:00+01:00 cv_wait_common+0x72/0xd0 [spl]
2012-01-08T18:29:00+01:00 ? add_wait_queue+0x60/0x60
2012-01-08T18:29:00+01:00 __cv_wait+0xe/0x10 [spl]
2012-01-08T18:29:00+01:00 txg_wait_open+0x73/0xa0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_wait+0xdd/0xe0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_assign+0x5e/0x410 [zfs]
2012-01-08T18:29:00+01:00 ? dsl_dataset_block_freeable+0x3b/0x50 [zfs]
2012-01-08T18:29:00+01:00 zvol_write+0x7c/0x380 [zfs]
2012-01-08T18:29:00+01:00 taskq_thread+0x1e2/0x410 [spl]
2012-01-08T18:29:00+01:00 ? try_to_wake_up+0x290/0x290
2012-01-08T18:29:00+01:00 ? task_alloc+0x160/0x160 [spl]
2012-01-08T18:29:00+01:00 kthread+0x87/0x90
2012-01-08T18:29:00+01:00 kernel_thread_helper+0x4/0x10
2012-01-08T18:29:00+01:00 ? flush_kthread_worker+0xa0/0xa0
2012-01-08T18:29:00+01:00 ? gs_change+0x13/0x13
2012-01-08T18:29:00+01:00          D
2012-01-08T18:29:00+01:00  1983      2 0x00000000
2012-01-08T18:29:00+01:00 Trace:
2012-01-08T18:29:00+01:00 schedule+0x3a/0x50
2012-01-08T18:29:00+01:00 cv_wait_common+0x72/0xd0 [spl]
2012-01-08T18:29:00+01:00 ? add_wait_queue+0x60/0x60
2012-01-08T18:29:00+01:00 __cv_wait+0xe/0x10 [spl]
2012-01-08T18:29:00+01:00 txg_wait_open+0x73/0xa0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_wait+0xdd/0xe0 [zfs]
2012-01-08T18:29:00+01:00 dmu_tx_assign+0x5e/0x410 [zfs]
2012-01-08T18:29:00+01:00 ? dsl_dataset_block_freeable+0x3b/0x50 [zfs]
2012-01-08T18:29:00+01:00 zvol_write+0x7c/0x380 [zfs]
2012-01-08T18:29:00+01:00 taskq_thread+0x1e2/0x410 [spl]
2012-01-08T18:29:00+01:00 ? try_to_wake_up+0x290/0x290
2012-01-08T18:29:00+01:00 ? task_alloc+0x160/0x160 [spl]
2012-01-08T18:29:00+01:00 kthread+0x87/0x90
2012-01-08T18:29:00+01:00 kernel_thread_helper+0x4/0x10
2012-01-08T18:29:00+01:00 ? flush_kthread_worker+0xa0/0xa0
2012-01-08T18:29:00+01:00 ? gs_change+0x13/0x13

kern: kern.info: Jan  8 18:30:38 kernel: sshd            D ffff88021e8203a8     0 31404  16413 0x00020000
kern: kern.notice: Jan  8 18:30:38 kernel:  ffff88027158d798 0000000000200086 0000000000000000 ffff8803ee6b5000
kern: kern.notice: Jan  8 18:30:38 kernel:  ffff88021e820000 ffff88027158dfd8 ffff88027158dfd8 ffff88027158dfd8
kern: kern.notice: Jan  8 18:30:38 kernel:  ffff8804054d5b40 ffff88021e820000 0000000000200282 00000001bf85b2c0
kern: kern.notice: Jan  8 18:30:38 kernel: Call Trace:
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff815e0dca>] schedule+0x3a/0x50
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa0326983>] zio_wait+0xe3/0x150 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa029fac7>] dbuf_read+0x337/0x740 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa02a636b>] dmu_buf_hold+0xbb/0x170 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa02fab5f>] zap_lockdir+0x4f/0x660 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff81304414>] ? timerqueue_del+0x34/0x90
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa02fc425>] zap_lookup_norm+0x45/0x190 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa02fc5ee>] zap_lookup+0x2e/0x30 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa0300f04>] zfs_dirent_lock+0x444/0x4f0 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa0301018>] zfs_dirlook+0x68/0x210 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa02ff6ce>] ? zfs_zaccess+0x9e/0x210 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa0318ea1>] zfs_lookup+0x2f1/0x340 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffffa032afd2>] zpl_lookup+0x52/0xb0 [zfs]
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff815e2999>] ? _raw_spin_lock+0x9/0x10
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff81186c39>] d_alloc_and_lookup+0x39/0x80
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff81188135>] do_lookup+0x2b5/0x3b0
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff811894d4>] path_lookupat+0x114/0x730
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff81189b1c>] do_path_lookup+0x2c/0xc0
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff8118b224>] user_path_at_empty+0x54/0xa0
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff812c3d51>] ? security_prepare_creds+0x11/0x20
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff8118b27c>] user_path_at+0xc/0x10
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff8117b04d>] sys_faccessat+0x9d/0x1e0
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff8117b1a3>] sys_access+0x13/0x20
kern: kern.notice: Jan  8 18:30:38 kernel:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13

So it does look similar to #496 after all.

The box isn't completely locked up (an i/o operation or two does complete every once in a while), but it's so slow it might as well be.

@akorn
Copy link
Contributor Author

akorn commented Jan 8, 2012

Well, it turns out the box was able to write at least parts of the kernel logs to disk (a non-zfs volume) during the deadlock, so I have some more process dump output to share:

18:30:54.443210500 kern.info: sshd            D ffff88021e8203a8     0 31404  16413 0x00020000
18:30:54.443280500 kern.notice:  ffff88027158d798 0000000000200086 0000000000000000 ffff8803ee6b5000
18:30:54.443353500 kern.notice:  ffff88021e820000 ffff88027158dfd8 ffff88027158dfd8 ffff88027158dfd8
18:30:54.443425500 kern.notice:  ffff8804054d5b40 ffff88021e820000 0000000000200282 00000001bf85b2c0
18:30:54.443497500 kern.notice: Call Trace:
18:30:54.443561500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.443652500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.443722500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.443793500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.443863500 kern.notice:  [<ffffffffa0326983>] zio_wait+0xe3/0x150 [zfs]
18:30:54.443932500 kern.notice:  [<ffffffffa029fac7>] dbuf_read+0x337/0x740 [zfs]
18:30:54.444000500 kern.notice:  [<ffffffffa02a636b>] dmu_buf_hold+0xbb/0x170 [zfs]
18:30:54.444070500 kern.notice:  [<ffffffffa02fab5f>] zap_lockdir+0x4f/0x660 [zfs]
18:30:54.444140500 kern.notice:  [<ffffffff81304414>] ? timerqueue_del+0x34/0x90
18:30:54.444209500 kern.notice:  [<ffffffffa02fc425>] zap_lookup_norm+0x45/0x190 [zfs]
18:30:54.444279500 kern.notice:  [<ffffffffa02fc5ee>] zap_lookup+0x2e/0x30 [zfs]
18:30:54.444348500 kern.notice:  [<ffffffffa0300f04>] zfs_dirent_lock+0x444/0x4f0 [zfs]
18:30:54.444427500 kern.notice:  [<ffffffffa0301018>] zfs_dirlook+0x68/0x210 [zfs]
18:30:54.444496500 kern.notice:  [<ffffffffa02ff6ce>] ? zfs_zaccess+0x9e/0x210 [zfs]
18:30:54.444566500 kern.notice:  [<ffffffffa0318ea1>] zfs_lookup+0x2f1/0x340 [zfs]
18:30:54.444633500 kern.notice:  [<ffffffffa032afd2>] zpl_lookup+0x52/0xb0 [zfs]
18:30:54.444708500 kern.notice:  [<ffffffff815e2999>] ? _raw_spin_lock+0x9/0x10
18:30:54.447959500 kern.notice:  [<ffffffff81186c39>] d_alloc_and_lookup+0x39/0x80
18:30:54.448032500 kern.notice:  [<ffffffff81188135>] do_lookup+0x2b5/0x3b0
18:30:54.448101500 kern.notice:  [<ffffffff811894d4>] path_lookupat+0x114/0x730
18:30:54.448169500 kern.notice:  [<ffffffff81189b1c>] do_path_lookup+0x2c/0xc0
18:30:54.448237500 kern.notice:  [<ffffffff8118b224>] user_path_at_empty+0x54/0xa0
18:30:54.448305500 kern.notice:  [<ffffffff812c3d51>] ? security_prepare_creds+0x11/0x20
18:30:54.448372500 kern.notice:  [<ffffffff8118b27c>] user_path_at+0xc/0x10
18:30:54.448471500 kern.notice:  [<ffffffff8117b04d>] sys_faccessat+0x9d/0x1e0
18:30:54.448542500 kern.notice:  [<ffffffff8117b1a3>] sys_access+0x13/0x20
18:30:54.448610500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.448677500 kern.info: ssh-agent       S ffff8801c3c583a8     0 31538      1 0x00020000
18:30:54.448747500 kern.notice:  ffff8802aa60b908 0000000000200082 006a5700010e7015 ff500000008c868d
18:30:54.448821500 kern.notice:  ffff8801c3c58000 ffff8802aa60bfd8 ffff8802aa60bfd8 ffff8802aa60bfd8
18:30:54.448895500 kern.notice:  ffff8804054ec470 ffff8801c3c58000 0000010168000000 4c15ff3f89047f89
18:30:54.448969500 kern.notice: Call Trace:
18:30:54.449033500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.449465500 kern.notice:  [<ffffffff815e1d95>] schedule_hrtimeout_range_clock+0x115/0x130
18:30:54.449538500 kern.notice:  [<ffffffff810b6078>] ? add_wait_queue+0x48/0x60
18:30:54.449609500 kern.notice:  [<ffffffff8118f39f>] ? __pollwait+0x6f/0xf0
18:30:54.449698500 kern.notice:  [<ffffffff815e1dbe>] schedule_hrtimeout_range+0xe/0x10
18:30:54.449769500 kern.notice:  [<ffffffff8118f244>] poll_schedule_timeout+0x44/0x60
18:30:54.449842500 kern.notice:  [<ffffffff8118fb8e>] do_select+0x56e/0x6b0
18:30:54.449913500 kern.notice:  [<ffffffff8113e371>] ? zone_statistics+0x91/0xb0
18:30:54.449983500 kern.notice:  [<ffffffff8118f330>] ? poll_freewait+0xd0/0xd0
18:30:54.450054500 kern.notice:  [<ffffffff8118f420>] ? __pollwait+0xf0/0xf0
18:30:54.450124500 kern.notice:  [<ffffffff8118f420>] ? __pollwait+0xf0/0xf0
18:30:54.450193500 kern.notice:  [<ffffffff8112a4d3>] ? __alloc_pages_nodemask+0x103/0x800
18:30:54.450270500 kern.notice:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
18:30:54.450342500 kern.notice:  [<ffffffff814cc8a1>] ? skb_dequeue+0x61/0x80
18:30:54.450410500 kern.notice:  [<ffffffff813019ce>] ? radix_tree_lookup_slot+0xe/0x10
18:30:54.450490500 kern.notice:  [<ffffffff81122409>] ? find_get_page+0x19/0x90
18:30:54.450560500 kern.notice:  [<ffffffff8112408f>] ? filemap_fault+0x7f/0x460
18:30:54.450630500 kern.notice:  [<ffffffff81121ab5>] ? unlock_page+0x25/0x30
18:30:54.450698500 kern.notice:  [<ffffffff81142de1>] ? __do_fault+0x3f1/0x4f0
18:30:54.450767500 kern.notice:  [<ffffffff811458df>] ? handle_pte_fault+0xef/0xab0
18:30:54.450836500 kern.notice:  [<ffffffff811c5960>] compat_core_sys_select+0x1c0/0x270
18:30:54.450907500 kern.notice:  [<ffffffff81146630>] ? handle_mm_fault+0x1f0/0x340
18:30:54.450976500 kern.notice:  [<ffffffff81072f10>] ? do_page_fault+0x1f0/0x4a0
18:30:54.451047500 kern.notice:  [<ffffffff81191755>] ? d_free+0x55/0x60
18:30:54.451113500 kern.notice:  [<ffffffff81199099>] ? vfsmount_lock_local_unlock+0x19/0x20
18:30:54.451184500 kern.notice:  [<ffffffff8119abeb>] ? mntput_no_expire+0x2b/0xe0
18:30:54.451264500 kern.notice:  [<ffffffff811c5ac0>] compat_sys_select+0xb0/0xf0
18:30:54.451346500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.451415500 kern.info: sshd            x ffff8803eb9c1a78     0 31544  31404 0x00020000
18:30:54.451483500 kern.notice:  ffff8802f49dde98 0000000000200046 0000000000000000 0000000000000004
18:30:54.451556500 kern.notice:  ffff8803eb9c16d0 ffff8802f49ddfd8 ffff8802f49ddfd8 ffff8802f49ddfd8
18:30:54.451642500 kern.notice:  ffff88033b258000 ffff8803eb9c16d0 ffff8803eb9c1cac ffff8803eb9c16d0
18:30:54.451715500 kern.notice: Call Trace:
18:30:54.451777500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.451928500 kern.notice:  [<ffffffff81098708>] do_exit+0x5b8/0x8b0
18:30:54.452076500 kern.notice:  [<ffffffff81098d3f>] do_group_exit+0x3f/0xa0
18:30:54.452145500 kern.notice:  [<ffffffff81098db2>] sys_exit_group+0x12/0x20
18:30:54.452213500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.452282500 kern.info: bash            S ffff88022a2503a8     0 31552      1 0x00020000
18:30:54.452356500 kern.notice:  ffff8801ce145d68 0000000000000082 ffff8801ce145cf8 ffffffff812c3f71
18:30:54.452428500 kern.notice:  ffff88022a250000 ffff8801ce145fd8 ffff8801ce145fd8 ffff8801ce145fd8
18:30:54.452501500 kern.notice:  ffff8804050b96d0 ffff88022a250000 ffff8801ce145d38 ffffffff8106f549
18:30:54.452574500 kern.notice: Call Trace:
18:30:54.452636500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.452706500 kern.notice:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
18:30:54.452775500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.452857500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.452928500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.452996500 kern.notice:  [<ffffffff81391349>] ? tty_ldisc_deref+0x9/0x10
18:30:54.453065500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.453134500 kern.notice:  [<ffffffff810e02c7>] compat_sys_wait4+0xc7/0xd0
18:30:54.453203500 kern.notice:  [<ffffffff810a91ed>] ? set_current_blocked+0x4d/0x60
18:30:54.453271500 kern.notice:  [<ffffffff810a9245>] ? sigprocmask+0x45/0x80
18:30:54.453339500 kern.notice:  [<ffffffff810a9303>] ? sys_rt_sigprocmask+0x83/0xa0
18:30:54.453408500 kern.notice:  [<ffffffff81078d4b>] ? sys32_rt_sigprocmask+0xdb/0x110
18:30:54.453477500 kern.notice:  [<ffffffff81078d9b>] sys32_waitpid+0xb/0x10
18:30:54.453544500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.453612500 kern.info: mc              D ffff8802ec85b148     0  1577  31552 0x00020000
18:30:54.453694500 kern.notice:  ffff8803f93d1a98 0000000000000082 ffff88029f243180 ffff8803e4bb3da8
18:30:54.453762500 kern.notice:  ffff8802ec85ada0 ffff8803f93d1fd8 ffff8803f93d1fd8 ffff8803f93d1fd8
18:30:54.453843500 kern.notice:  ffff8804054d5b40 ffff8802ec85ada0 0000000000000282 ffff880402ea7b30
18:30:54.453918500 kern.notice: Call Trace:
18:30:54.453980500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.454048500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.454114500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.454183500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.454252500 kern.notice:  [<ffffffffa02e7d33>] txg_wait_open+0x73/0xa0 [zfs]
18:30:54.454323500 kern.notice:  [<ffffffffa02b125d>] dmu_tx_wait+0xdd/0xe0 [zfs]
18:30:54.454392500 kern.notice:  [<ffffffffa0318670>] zfs_create+0x2a0/0x6c0 [zfs]
18:30:54.454465500 kern.notice:  [<ffffffff815e2999>] ? _raw_spin_lock+0x9/0x10
18:30:54.454543500 kern.notice:  [<ffffffff8119207a>] ? __d_instantiate+0x8a/0x110
18:30:54.454613500 kern.notice:  [<ffffffffa032b242>] zpl_create+0xa2/0xe0 [zfs]
18:30:54.454681500 kern.notice:  [<ffffffff8118af6c>] vfs_create+0xac/0x110
18:30:54.454749500 kern.notice:  [<ffffffff8118c820>] do_last+0x840/0x9b0
18:30:54.454819500 kern.notice:  [<ffffffff8118ca6f>] path_openat+0xdf/0x490
18:30:54.454886500 kern.notice:  [<ffffffffa0319372>] ? zfs_read+0x3b2/0x480 [zfs]
18:30:54.454955500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.455023500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.455090500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.455158500 kern.notice:  [<ffffffff8117a561>] ? filp_close+0x61/0x90
18:30:54.455226500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.455293500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.455371500 kern.info: cron            S ffff8801cb389a78     0  3598   5928 0x00000000
18:30:54.455442500 kern.notice:  ffff8801e6885e58 0000000000000082 ffff880100000000 ffffffff812c3f71
18:30:54.455514500 kern.notice:  ffff8801cb3896d0 ffff8801e6885fd8 ffff8801e6885fd8 ffff8801e6885fd8
18:30:54.455586500 kern.notice:  ffff8804054ec470 ffff8801cb3896d0 ffff8801e6885e28 000000018106f549
18:30:54.455658500 kern.notice: Call Trace:
18:30:54.455720500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.455789500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.455856500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.455923500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.455991500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.456059500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.456128500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.456210500 kern.info: sh              S ffff8801c93f9a78     0  3604   3598 0x00000000
18:30:54.456289500 kern.notice:  ffff880267cffe58 0000000000000082 ffff880200000000 ffffffff812c3f71
18:30:54.456364500 kern.notice:  ffff8801c93f96d0 ffff880267cfffd8 ffff880267cfffd8 ffff880267cfffd8
18:30:54.456441500 kern.notice:  ffff8804054d5b40 ffff8801c93f96d0 ffff880267cffe28 000000018106f549
18:30:54.456514500 kern.notice: Call Trace:
18:30:54.456578500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.456647500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.456715500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.456783500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.456850500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.456918500 kern.notice:  [<ffffffff811812f5>] ? sys_newstat+0x25/0x30
18:30:54.456987500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.457065500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.457134500 kern.info: zfs-auto-snapsh S ffff8801c8de1a78     0  3605   3604 0x00000000
18:30:54.457200500 kern.notice:  ffff880314587e58 0000000000000082 ffff880314587de8 ffffffff812c3f71
18:30:54.457273500 kern.notice:  ffff8801c8de16d0 ffff880314587fd8 ffff880314587fd8 ffff880314587fd8
18:30:54.457345500 kern.notice:  ffff8802921ac470 ffff8801c8de16d0 ffff880314587e28 ffffffff8106f549
18:30:54.457417500 kern.notice: Call Trace:
18:30:54.457479500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.457548500 kern.notice:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
18:30:54.457617500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.457685500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.457753500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.457832500 kern.notice:  [<ffffffff810824f4>] ? finish_task_switch+0x44/0xe0
18:30:54.457901500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.457969500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.458038500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.458107500 kern.info: cron            S ffff8803f34e83a8     0  5317   5928 0x00000000
18:30:54.458179500 kern.notice:  ffff8801bc983e58 0000000000000086 ffff880100000000 ffffffff812c3f71
18:30:54.458252500 kern.notice:  ffff8803f34e8000 ffff8801bc983fd8 ffff8801bc983fd8 ffff8801bc983fd8
18:30:54.458324500 kern.notice:  ffff8804054d0000 ffff8803f34e8000 ffff8801bc983e28 000000018106f549
18:30:54.458401500 kern.notice: Call Trace:
18:30:54.458464500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.458534500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.458601500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.458688500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.458759500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.458826500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.458895500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.458964500 kern.info: sh              S ffff8803f34ec818     0  5321   5317 0x00000000
18:30:54.459033500 kern.notice:  ffff88035ff63e58 0000000000000082 ffff880300000000 ffffffff812c3f71
18:30:54.459115500 kern.notice:  ffff8803f34ec470 ffff88035ff63fd8 ffff88035ff63fd8 ffff88035ff63fd8
18:30:54.459139500 kern.notice:  ffff8804054d0000 ffff8803f34ec470 ffff88035ff63e28 000000018106f549
18:30:54.459162500 kern.notice: Call Trace:
18:30:54.459182500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.459204500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.459306500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.459328500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.459350500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.459372500 kern.notice:  [<ffffffff811812f5>] ? sys_newstat+0x25/0x30
18:30:54.459393500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.459415500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.459437500 kern.info: run-parts       S ffff8803e8764818     0  5322   5321 0x00000000
18:30:54.459459500 kern.notice:  ffff8803207d9908 0000000000000082 ffff8803207d98c8 ffffffff8117434f
18:30:54.459481500 kern.notice:  ffff8803e8764470 ffff8803207d9fd8 ffff8803207d9fd8 ffff8803207d9fd8
18:30:54.459505500 kern.notice:  ffff880402064470 ffff8803e8764470 ffff8803207d98e8 ffffffff81122409
18:30:54.459528500 kern.notice: Call Trace:
18:30:54.459548500 kern.notice:  [<ffffffff8117434f>] ? __mem_cgroup_commit_charge+0x6f/0xd0
18:30:54.459574500 kern.notice:  [<ffffffff81122409>] ? find_get_page+0x19/0x90
18:30:54.459597500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.459619500 kern.notice:  [<ffffffff815e1d95>] schedule_hrtimeout_range_clock+0x115/0x130
18:30:54.459641500 kern.notice:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
18:30:54.459662500 kern.notice:  [<ffffffff810b6078>] ? add_wait_queue+0x48/0x60
18:30:54.459684500 kern.notice:  [<ffffffff815e1dbe>] schedule_hrtimeout_range+0xe/0x10
18:30:54.459706500 kern.notice:  [<ffffffff8118f244>] poll_schedule_timeout+0x44/0x60
18:30:54.459728500 kern.notice:  [<ffffffff8118fb8e>] do_select+0x56e/0x6b0
18:30:54.459749500 kern.notice:  [<ffffffff8113e371>] ? zone_statistics+0x91/0xb0
18:30:54.459772500 kern.notice:  [<ffffffff8118f330>] ? poll_freewait+0xd0/0xd0
18:30:54.459793500 kern.notice:  [<ffffffff8118f420>] ? __pollwait+0xf0/0xf0
18:30:54.459822500 kern.notice:  [<ffffffff8118f420>] ? __pollwait+0xf0/0xf0
18:30:54.459844500 kern.notice:  [<ffffffff8112dd90>] ? pagevec_move_tail+0x30/0x30
18:30:54.459866500 kern.notice:  [<ffffffff8112dd21>] ? pagevec_lru_move_fn+0xd1/0xf0
18:30:54.459888500 kern.notice:  [<ffffffff8112dd52>] ? ____pagevec_lru_add+0x12/0x20
18:30:54.459910500 kern.notice:  [<ffffffff8112dfc8>] ? __lru_cache_add+0x68/0x90
18:30:54.459932500 kern.notice:  [<ffffffff81077ce3>] ? flush_tlb_page+0x43/0xa0
18:30:54.459954500 kern.notice:  [<ffffffff8112e358>] ? lru_cache_add_lru+0x28/0x50
18:30:54.459976500 kern.notice:  [<ffffffff8114f615>] ? page_add_new_anon_rmap+0x95/0xa0
18:30:54.459998500 kern.notice:  [<ffffffff81143fb8>] ? do_wp_page+0x2d8/0x720
18:30:54.460020500 kern.notice:  [<ffffffff81147235>] ? copy_page_range+0x255/0x4b0
18:30:54.460042500 kern.notice:  [<ffffffff81145a75>] ? handle_pte_fault+0x285/0xab0
18:30:54.460064500 kern.notice:  [<ffffffff8118fe85>] core_sys_select+0x1b5/0x320
18:30:54.460089500 kern.notice:  [<ffffffff81146630>] ? handle_mm_fault+0x1f0/0x340
18:30:54.460111500 kern.notice:  [<ffffffff81066301>] ? native_smp_send_reschedule+0x41/0x60
18:30:54.460133500 kern.notice:  [<ffffffff8107ca2e>] ? resched_task+0x6e/0x80
18:30:54.460155500 kern.notice:  [<ffffffff81072f10>] ? do_page_fault+0x1f0/0x4a0
18:30:54.460176500 kern.notice:  [<ffffffff8108c9d5>] ? wake_up_new_task+0xf5/0x180
18:30:54.460198500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.460221500 kern.notice:  [<ffffffff811900a6>] sys_select+0xb6/0x100
18:30:54.460244500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.460266500 kern.info: zfs-auto-snapsh S ffff8803f3464818     0  5323   5322 0x00000000
18:30:54.460287500 kern.notice:  ffff88027a0a9e58 0000000000000082 ffff88027a0a9de8 ffffffff812c3f71
18:30:54.460309500 kern.notice:  ffff8803f3464470 ffff88027a0a9fd8 ffff88027a0a9fd8 ffff88027a0a9fd8
18:30:54.460338500 kern.notice:  ffff8803ecfd96d0 ffff8803f3464470 ffff88027a0a9e28 ffffffff8106f549
18:30:54.460361500 kern.notice: Call Trace:
18:30:54.460381500 kern.notice:  [<ffffffff812c3f71>] ? security_task_wait+0x11/0x20
18:30:54.460403500 kern.notice:  [<ffffffff8106f549>] ? default_spin_lock_flags+0x9/0x10
18:30:54.460425500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.460446500 kern.notice:  [<ffffffff81097a5e>] do_wait+0x1de/0x260
18:30:54.460467500 kern.notice:  [<ffffffff81093952>] ? do_fork+0x182/0x310
18:30:54.460489500 kern.notice:  [<ffffffff810824f4>] ? finish_task_switch+0x44/0xe0
18:30:54.460511500 kern.notice:  [<ffffffff81098fee>] sys_wait4+0x8e/0xe0
18:30:54.460533500 kern.notice:  [<ffffffff8104c3ab>] ? sys_rt_sigreturn+0x1eb/0x200
18:30:54.460555500 kern.notice:  [<ffffffff81096780>] ? task_stopped_code+0x40/0x40
18:30:54.460581500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.460607500 kern.info: kworker/0:0     S ffff880290529a78     0  5494      2 0x00000000
18:30:54.460629500 kern.notice:  ffff8802f47c9e60 0000000000000046 ffff8802f47c9e00 ffffffff8124ba3e
18:30:54.460652500 kern.notice:  ffff8802905296d0 ffff8802f47c9fd8 ffff8802f47c9fd8 ffff8802f47c9fd8
18:30:54.460675500 kern.notice:  ffff880233340000 ffff8802905296d0 ffffffffffffff02 ffffffff815e0488
18:30:54.460699500 kern.notice: Call Trace:
18:30:54.460719500 kern.notice:  [<ffffffff8124ba3e>] ? xfs_buf_iodone_work+0x1e/0x50
18:30:54.460740500 kern.notice:  [<ffffffff815e0488>] ? __schedule+0x18/0x890
18:30:54.460762500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.460783500 kern.notice:  [<ffffffff810b15ce>] worker_thread+0x22e/0x340
18:30:54.460805500 kern.notice:  [<ffffffff810b13a0>] ? manage_workers.isra.31+0x230/0x230
18:30:54.460827500 kern.notice:  [<ffffffff810b5667>] kthread+0x87/0x90
18:30:54.460850500 kern.notice:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
18:30:54.460876500 kern.notice:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
18:30:54.460898500 kern.notice:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
18:30:54.460919500 kern.info: kworker/4:1     S ffff88021d921a78     0  5952      2 0x00000000
18:30:54.460942500 kern.notice:  ffff8803f79f9e60 0000000000000046 ffff8803f79f9e00 ffffffff812e54d9
18:30:54.460965500 kern.notice:  ffff88021d9216d0 ffff8803f79f9fd8 ffff8803f79f9fd8 ffff8803f79f9fd8
18:30:54.460987500 kern.notice:  ffff880405508000 ffff88021d9216d0 0000000000000003 0000000100000001
18:30:54.461010500 kern.notice: Call Trace:
18:30:54.461030500 kern.notice:  [<ffffffff812e54d9>] ? blk_delay_work+0x39/0x50
18:30:54.461051500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.461073500 kern.notice:  [<ffffffff810b15ce>] worker_thread+0x22e/0x340
18:30:54.461094500 kern.notice:  [<ffffffff810b13a0>] ? manage_workers.isra.31+0x230/0x230
18:30:54.461120500 kern.notice:  [<ffffffff810b5667>] kthread+0x87/0x90
18:30:54.461142500 kern.notice:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
18:30:54.461164500 kern.notice:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
18:30:54.461186500 kern.notice:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
18:30:54.461207500 kern.info: kworker/0:1     R  running task        0  6151      2 0x00000000
18:30:54.461230500 kern.notice:  ffff88031ee35e60 0000000000000046 ffff88031ee35e00 ffffffffa00ca3f8
18:30:54.461252500 kern.notice:  ffff880310ae0000 ffff88031ee35fd8 ffff88031ee35fd8 ffff88031ee35fd8
18:30:54.461276500 kern.notice:  ffff8803fb4a2da0 ffff880310ae0000 ffff880405279300 00000001af4a4f00
18:30:54.461299500 kern.notice: Call Trace:
18:30:54.461319500 kern.notice:  [<ffffffffa00ca3f8>] ? spl_cache_age+0x48/0x50 [spl]
18:30:54.461344500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.461365500 kern.notice:  [<ffffffff810b15ce>] worker_thread+0x22e/0x340
18:30:54.461390500 kern.notice:  [<ffffffff810b13a0>] ? manage_workers.isra.31+0x230/0x230
18:30:54.461412500 kern.notice:  [<ffffffff810b5667>] kthread+0x87/0x90
18:30:54.461434500 kern.notice:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
18:30:54.461456500 kern.notice:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
18:30:54.461480500 kern.notice:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
18:30:54.461501500 kern.info: kworker/0:5     S ffff8803055583a8     0  6569      2 0x00000000
18:30:54.461524500 kern.notice:  ffff8803a7f9be60 0000000000000046 ffff8803a7f9be00 ffffffff812e54d9
18:30:54.461547500 kern.notice:  ffff880305558000 ffff8803a7f9bfd8 ffff8803a7f9bfd8 ffff8803a7f9bfd8
18:30:54.461570500 kern.notice:  ffff880310ae0000 ffff880305558000 ffff880405279300 0000000105558000
18:30:54.461593500 kern.notice: Call Trace:
18:30:54.461613500 kern.notice:  [<ffffffff812e54d9>] ? blk_delay_work+0x39/0x50
18:30:54.461638500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.461659500 kern.notice:  [<ffffffff810b15ce>] worker_thread+0x22e/0x340
18:30:54.461681500 kern.notice:  [<ffffffff810b13a0>] ? manage_workers.isra.31+0x230/0x230
18:30:54.461703500 kern.notice:  [<ffffffff810b5667>] kthread+0x87/0x90
18:30:54.461725500 kern.notice:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
18:30:54.461747500 kern.notice:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
18:30:54.461770500 kern.notice:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
18:30:54.461791500 kern.info: zfs             D ffff8802921ac818     0  6794   3605 0x00000000
18:30:54.461813500 kern.notice:  ffff8802a7c71cc8 0000000000000082 ffffffff00000000 ffff8803e79d1400
18:30:54.461836500 kern.notice:  ffff8802921ac470 ffff8802a7c71fd8 ffff8802a7c71fd8 ffff8802a7c71fd8
18:30:54.461859500 kern.notice:  ffff88023d6ec470 ffff8802921ac470 0000000000000286 ffff880402ea7a80
18:30:54.461885500 kern.notice: Call Trace:
18:30:54.461905500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.461926500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.461948500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.461971500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.461992500 kern.notice:  [<ffffffffa02e7dd3>] txg_wait_synced+0x73/0xa0 [zfs]
18:30:54.462014500 kern.notice:  [<ffffffffa02cd979>] dsl_sync_task_group_wait+0x109/0x230 [zfs]
18:30:54.462036500 kern.notice:  [<ffffffffa02bb399>] dsl_snapshots_destroy+0x139/0x150 [zfs]
18:30:54.462058500 kern.notice:  [<ffffffffa0307f4e>] zfs_ioc_destroy_snaps+0x6e/0x80 [zfs]
18:30:54.462083500 kern.notice:  [<ffffffffa030be27>] zfsdev_ioctl+0xd7/0x1b0 [zfs]
18:30:54.462107500 kern.notice:  [<ffffffff8118e65e>] do_vfs_ioctl+0x8e/0x4e0
18:30:54.462128500 kern.notice:  [<ffffffff81167c6a>] ? kmem_cache_free+0x2a/0x100
18:30:54.462185500 kern.notice:  [<ffffffff8118732e>] ? putname+0x2e/0x40
18:30:54.462207500 kern.notice:  [<ffffffff8118eb39>] sys_ioctl+0x89/0x90
18:30:54.462228500 kern.notice:  [<ffffffff8104d5a9>] ? do_device_not_available+0x9/0x10
18:30:54.462251500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.462272500 kern.info: zpool           S ffff88023d6e9a78     0  9282  22845 0x00000000
18:30:54.462294500 kern.notice:  ffff8801b6361e78 0000000000000082 00000000b2d05e00 0000000100000000
18:30:54.462317500 kern.notice:  ffff88023d6e96d0 ffff8801b6361fd8 ffff8801b6361fd8 ffff8801b6361fd8
18:30:54.462341500 kern.notice:  ffff880405508000 ffff88023d6e96d0 ffff880100000001 0000000000000282
18:30:54.462364500 kern.notice: Call Trace:
18:30:54.462383500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.462405500 kern.notice:  [<ffffffff815e1bd4>] do_nanosleep+0x94/0xd0
18:30:54.462426500 kern.notice:  [<ffffffff810ba4eb>] hrtimer_nanosleep+0x9b/0x140
18:30:54.462451500 kern.notice:  [<ffffffff810b9310>] ? update_rmtp+0x60/0x60
18:30:54.462473500 kern.notice:  [<ffffffff810ba5e2>] sys_nanosleep+0x52/0x70
18:30:54.462494500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.462516500 kern.info: zfs             D ffff88023d6ec818     0  9285   5323 0x00000000
18:30:54.462538500 kern.notice:  ffff8801c78b9c58 0000000000000086 ffff8804051b6400 0000000000000000
18:30:54.462561500 kern.notice:  ffff88023d6ec470 ffff8801c78b9fd8 ffff8801c78b9fd8 ffff8801c78b9fd8
18:30:54.462584500 kern.notice:  ffff880394e2c470 ffff88023d6ec470 0000000000000286 ffff880402ea7a80
18:30:54.462607500 kern.notice: Call Trace:
18:30:54.462626500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.462647500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.462670500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.462695500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.462717500 kern.notice:  [<ffffffffa02e7dd3>] txg_wait_synced+0x73/0xa0 [zfs]
18:30:54.462738500 kern.notice:  [<ffffffffa02cd979>] dsl_sync_task_group_wait+0x109/0x230 [zfs]
18:30:54.462760500 kern.notice:  [<ffffffffa02ab15a>] dmu_objset_snapshot+0x2aa/0x2f0 [zfs]
18:30:54.462784500 kern.notice:  [<ffffffffa02df17e>] ? spa_open+0xe/0x10 [zfs]
18:30:54.462806500 kern.notice:  [<ffffffffa03094ac>] zfs_ioc_snapshot+0x1bc/0x200 [zfs]
18:30:54.462827500 kern.notice:  [<ffffffffa030be27>] zfsdev_ioctl+0xd7/0x1b0 [zfs]
18:30:54.462849500 kern.notice:  [<ffffffff8118e65e>] do_vfs_ioctl+0x8e/0x4e0
18:30:54.462874500 kern.notice:  [<ffffffff81167c6a>] ? kmem_cache_free+0x2a/0x100
18:30:54.462897500 kern.notice:  [<ffffffff8118732e>] ? putname+0x2e/0x40
18:30:54.462917500 kern.notice:  [<ffffffff8118eb39>] sys_ioctl+0x89/0x90
18:30:54.462938500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.462964500 kern.info: top             S ffff8802ec85c818     0  9306  19159 0x00000000
18:30:54.462985500 kern.notice:  ffff8801b470bde8 0000000000000082 ffff880100000000 ffffffff8107cd26
18:30:54.463008500 kern.notice:  ffff8802ec85c470 ffff8801b470bfd8 ffff8801b470bfd8 ffff8801b470bfd8
18:30:54.463031500 kern.notice:  ffff8804054d0000 ffff8802ec85c470 ffff8801b470bdc8 0000000181388ad9
18:30:54.463054500 kern.notice: Call Trace:
18:30:54.463074500 kern.notice:  [<ffffffff8107cd26>] ? __wake_up+0x46/0x60
18:30:54.463096500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.463118500 kern.notice:  [<ffffffff8138d4a6>] n_tty_write+0x226/0x450
18:30:54.463139500 kern.notice:  [<ffffffff8108c460>] ? try_to_wake_up+0x290/0x290
18:30:54.463162500 kern.notice:  [<ffffffff81389ec9>] tty_write+0x149/0x290
18:30:54.463184500 kern.notice:  [<ffffffff8138d280>] ? process_echoes+0x320/0x320
18:30:54.463205500 kern.notice:  [<ffffffff8117c81e>] vfs_write+0xae/0x170
18:30:54.463233500 kern.notice:  [<ffffffff8117cb25>] sys_write+0x45/0x90
18:30:54.463254500 kern.notice:  [<ffffffff815e35d2>] system_call_fastpath+0x16/0x1b
18:30:54.463276500 kern.info: cron            D ffff88023d6edee8     0  9310  16501 0x00020000
18:30:54.463299500 kern.notice:  ffff8803e1ee5a98 0000000000000086 ffff8802ee2a13c0 ffff8803407c3a90
18:30:54.463322500 kern.notice:  ffff88023d6edb40 ffff8803e1ee5fd8 ffff8803e1ee5fd8 ffff8803e1ee5fd8
18:30:54.463345500 kern.notice:  ffff8803ec935b40 ffff88023d6edb40 0000000000000282 ffff880402ea7b30
18:30:54.463368500 kern.notice: Call Trace:
18:30:54.463388500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.463409500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.463430500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.463454500 kern.notice:  [<ffffffffa00c9071>] ? kmem_free_debug+0x11/0x20 [spl]
18:30:54.463479500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.463501500 kern.notice:  [<ffffffffa02e7d33>] txg_wait_open+0x73/0xa0 [zfs]
18:30:54.463523500 kern.notice:  [<ffffffffa02b125d>] dmu_tx_wait+0xdd/0xe0 [zfs]
18:30:54.463545500 kern.notice:  [<ffffffffa0318670>] zfs_create+0x2a0/0x6c0 [zfs]
18:30:54.463566500 kern.notice:  [<ffffffff815e2999>] ? _raw_spin_lock+0x9/0x10
18:30:54.463588500 kern.notice:  [<ffffffff8119207a>] ? __d_instantiate+0x8a/0x110
18:30:54.463610500 kern.notice:  [<ffffffffa032b242>] zpl_create+0xa2/0xe0 [zfs]
18:30:54.463632500 kern.notice:  [<ffffffff8118af6c>] vfs_create+0xac/0x110
18:30:54.463657500 kern.notice:  [<ffffffff8118c820>] do_last+0x840/0x9b0
18:30:54.463679500 kern.notice:  [<ffffffff8118ca6f>] path_openat+0xdf/0x490
18:30:54.463700500 kern.notice:  [<ffffffffa00cf618>] ? tsd_exit+0xf8/0x110 [spl]
18:30:54.463722500 kern.notice:  [<ffffffff81146630>] ? handle_mm_fault+0x1f0/0x340
18:30:54.463747500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.463769500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.463791500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.463812500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.463834500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.463855500 kern.info: cron            D ffff88023d6e83a8     0  9311  16501 0x00020000
18:30:54.463877500 kern.notice:  ffff8802587a7c78 0000000000000086 ffff880100000000 0000000000000003
18:30:54.463900500 kern.notice:  ffff88023d6e8000 ffff8802587a7fd8 ffff8802587a7fd8 ffff8802587a7fd8
18:30:54.463923500 kern.notice:  ffff88040550ada0 ffff88023d6e8000 ffff8802587a7cb8 0000000181187f47
18:30:54.463946500 kern.notice: Call Trace:
18:30:54.463966500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.463987500 kern.notice:  [<ffffffff815e1ac7>] __mutex_lock_slowpath+0xc7/0x140
18:30:54.464012500 kern.notice:  [<ffffffff815e16cd>] mutex_lock+0x1d/0x40
18:30:54.464034500 kern.notice:  [<ffffffff8118c37b>] do_last+0x39b/0x9b0
18:30:54.464055500 kern.notice:  [<ffffffff8118ca6f>] path_openat+0xdf/0x490
18:30:54.464078500 kern.notice:  [<ffffffffa00cf618>] ? tsd_exit+0xf8/0x110 [spl]
18:30:54.464100500 kern.notice:  [<ffffffff81146630>] ? handle_mm_fault+0x1f0/0x340
18:30:54.464122500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.464144500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.464165500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.464187500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.464209500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.464230500 kern.info: cron            D ffff88023d6eb148     0  9312  16501 0x00020000
18:30:54.464252500 kern.notice:  ffff8803005d5c78 0000000000000082 ffff880200000000 0000000000000003
18:30:54.464280500 kern.notice:  ffff88023d6eada0 ffff8803005d5fd8 ffff8803005d5fd8 ffff8803005d5fd8
18:30:54.464303500 kern.notice:  ffff8804055216d0 ffff88023d6eada0 ffff8803005d5cb8 0000000181187f47
18:30:54.464324500 kern.notice: Call Trace:
18:30:54.464344500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.464366500 kern.notice:  [<ffffffff815e1ac7>] __mutex_lock_slowpath+0xc7/0x140
18:30:54.464388500 kern.notice:  [<ffffffff815e16cd>] mutex_lock+0x1d/0x40
18:30:54.464412500 kern.notice:  [<ffffffff8118c37b>] do_last+0x39b/0x9b0
18:30:54.464435500 kern.notice:  [<ffffffff8118ca6f>] path_openat+0xdf/0x490
18:30:54.464456500 kern.notice:  [<ffffffffa00cf618>] ? tsd_exit+0xf8/0x110 [spl]
18:30:54.464478500 kern.notice:  [<ffffffff81146630>] ? handle_mm_fault+0x1f0/0x340
18:30:54.464500500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.464521500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.464546500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.464568500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.464589500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.464611500 kern.info: munin-node      D ffff88031add83a8     0  9374  16407 0x00020000
18:30:54.464633500 kern.notice:  ffff88032ce91a98 0000000000000086 ffff880200000000 ffff8803e4bb3da8
18:30:54.464655500 kern.notice:  ffff88031add8000 ffff88032ce91fd8 ffff88032ce91fd8 ffff88032ce91fd8
18:30:54.464678500 kern.notice:  ffff88040550ada0 ffff88031add8000 0000000000000282 0000000102ea7b30
18:30:54.464701500 kern.notice: Call Trace:
18:30:54.464721500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.464742500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.464766500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.464788500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.464813500 kern.notice:  [<ffffffffa02e7d33>] txg_wait_open+0x73/0xa0 [zfs]
18:30:54.464834500 kern.notice:  [<ffffffffa02b125d>] dmu_tx_wait+0xdd/0xe0 [zfs]
18:30:54.464856500 kern.notice:  [<ffffffffa0318670>] zfs_create+0x2a0/0x6c0 [zfs]
18:30:54.464878500 kern.notice:  [<ffffffffa032b242>] zpl_create+0xa2/0xe0 [zfs]
18:30:54.464900500 kern.notice:  [<ffffffff8118af6c>] vfs_create+0xac/0x110
18:30:54.464921500 kern.notice:  [<ffffffff8118c820>] do_last+0x840/0x9b0
18:30:54.464943500 kern.notice:  [<ffffffff8118ca6f>] path_openat+0xdf/0x490
18:30:54.464965500 kern.notice:  [<ffffffff8118732e>] ? putname+0x2e/0x40
18:30:54.464987500 kern.notice:  [<ffffffff8118b232>] ? user_path_at_empty+0x62/0xa0
18:30:54.465008500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.465030500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.465052500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.465076500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.465098500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.465120500 kern.info: flush-zfs-7     S ffff8803f34603a8     0  9389      2 0x00000000
18:30:54.465142500 kern.notice:  ffff880229301e50 0000000000000046 ffff8803f169d130 0000000000000000
18:30:54.465165500 kern.notice:  ffff8803f3460000 ffff880229301fd8 ffff880229301fd8 ffff880229301fd8
18:30:54.465192500 kern.notice:  ffff8802c5445b40 ffff8803f3460000 ffffffff810a2c90 ffff8803f3460000
18:30:54.465216500 kern.notice: Call Trace:
18:30:54.465235500 kern.notice:  [<ffffffff810a2c90>] ? usleep_range+0x40/0x40
18:30:54.465257500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.465279500 kern.notice:  [<ffffffff811a58c0>] bdi_writeback_thread+0x180/0x290
18:30:54.465301500 kern.notice:  [<ffffffff811a5740>] ? wb_do_writeback+0x200/0x200
18:30:54.465330500 kern.notice:  [<ffffffff810b5667>] kthread+0x87/0x90
18:30:54.465352500 kern.notice:  [<ffffffff815e4d34>] kernel_thread_helper+0x4/0x10
18:30:54.465374500 kern.notice:  [<ffffffff810b55e0>] ? flush_kthread_worker+0xa0/0xa0
18:30:54.465397500 kern.notice:  [<ffffffff815e4d30>] ? gs_change+0x13/0x13
18:30:54.465420500 kern.info: cron            D ffff8803eb9c4818     0  9450  16501 0x00020000
18:30:54.465443500 kern.notice:  ffff88032cf3fc78 0000000000000082 ffff880200000000 0000000000000003
18:30:54.465466500 kern.notice:  ffff8803eb9c4470 ffff88032cf3ffd8 ffff88032cf3ffd8 ffff88032cf3ffd8
18:30:54.465489500 kern.notice:  ffff8804054ec470 ffff8803eb9c4470 ffff88032cf3fcb8 0000000181187f47
18:30:54.465512500 kern.notice: Call Trace:
18:30:54.465532500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.465553500 kern.notice:  [<ffffffff815e1ac7>] __mutex_lock_slowpath+0xc7/0x140
18:30:54.465575500 kern.notice:  [<ffffffff815e16cd>] mutex_lock+0x1d/0x40
18:30:54.465599500 kern.notice:  [<ffffffff8118c37b>] do_last+0x39b/0x9b0
18:30:54.465621500 kern.notice:  [<ffffffff8118ca6f>] path_openat+0xdf/0x490
18:30:54.465642500 kern.notice:  [<ffffffffa00cf618>] ? tsd_exit+0xf8/0x110 [spl]
18:30:54.465664500 kern.notice:  [<ffffffff81146630>] ? handle_mm_fault+0x1f0/0x340
18:30:54.465686500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.465707500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.465728500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.465750500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.465772500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.465793500 kern.info: sshd            D ffff88036bdfc818     0 11473  16413 0x00020000
18:30:54.465815500 kern.notice:  ffff8803285e9718 0000000000200082 ffff880300000000 ffff8803ee6b5000
18:30:54.465838500 kern.notice:  ffff88036bdfc470 ffff8803285e9fd8 ffff8803285e9fd8 ffff8803285e9fd8
18:30:54.465864500 kern.notice:  ffff8804054d0000 ffff88036bdfc470 0000000000200286 0000000166755b20
18:30:54.465887500 kern.notice: Call Trace:
18:30:54.465907500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.465929500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.465955500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.465978500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.466000500 kern.notice:  [<ffffffffa0326983>] zio_wait+0xe3/0x150 [zfs]
18:30:54.466022500 kern.notice:  [<ffffffffa029fac7>] dbuf_read+0x337/0x740 [zfs]
18:30:54.466045500 kern.notice:  [<ffffffffa02a636b>] dmu_buf_hold+0xbb/0x170 [zfs]
18:30:54.466069500 kern.notice:  [<ffffffffa02fab5f>] zap_lockdir+0x4f/0x660 [zfs]
18:30:54.466091500 kern.notice:  [<ffffffffa02fc425>] zap_lookup_norm+0x45/0x190 [zfs]
18:30:54.466113500 kern.notice:  [<ffffffffa02fc5ee>] zap_lookup+0x2e/0x30 [zfs]
18:30:54.466138500 kern.notice:  [<ffffffffa0300f04>] zfs_dirent_lock+0x444/0x4f0 [zfs]
18:30:54.466160500 kern.notice:  [<ffffffffa0301018>] zfs_dirlook+0x68/0x210 [zfs]
18:30:54.466182500 kern.notice:  [<ffffffffa02ff6ce>] ? zfs_zaccess+0x9e/0x210 [zfs]
18:30:54.466203500 kern.notice:  [<ffffffffa0318ea1>] zfs_lookup+0x2f1/0x340 [zfs]
18:30:54.466225500 kern.notice:  [<ffffffffa032afd2>] zpl_lookup+0x52/0xb0 [zfs]
18:30:54.466246500 kern.notice:  [<ffffffff815e2999>] ? _raw_spin_lock+0x9/0x10
18:30:54.466268500 kern.notice:  [<ffffffff81186c39>] d_alloc_and_lookup+0x39/0x80
18:30:54.466290500 kern.notice:  [<ffffffff81188135>] do_lookup+0x2b5/0x3b0
18:30:54.466312500 kern.notice:  [<ffffffff812c3887>] ? security_inode_permission+0x17/0x20
18:30:54.466334500 kern.notice:  [<ffffffff81188ba7>] link_path_walk+0x137/0x860
18:30:54.466354500 kern.notice:  [<ffffffff8119abeb>] ? mntput_no_expire+0x2b/0xe0
18:30:54.466379500 kern.notice:  [<ffffffff8118ccb0>] path_openat+0x320/0x490
18:30:54.466401500 kern.notice:  [<ffffffff8118cf2d>] do_filp_open+0x3d/0xa0
18:30:54.466422500 kern.notice:  [<ffffffff81198954>] ? alloc_fd+0x104/0x150
18:30:54.466444500 kern.notice:  [<ffffffff8117bcb2>] do_sys_open+0xf2/0x1d0
18:30:54.466467500 kern.notice:  [<ffffffff811c5785>] compat_sys_open+0x15/0x20
18:30:54.466489500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.466510500 kern.info: tcpsvd          D ffff8801b1411a78     0 12237  16425 0x00020000
18:30:54.466532500 kern.notice:  ffff8801b0203658 0000000000000086 ffff880100000000 ffff8803ee6b5000
18:30:54.466555500 kern.notice:  ffff8801b14116d0 ffff8801b0203fd8 ffff8801b0203fd8 ffff8801b0203fd8
18:30:54.466579500 kern.notice:  ffff88040550ada0 ffff8801b14116d0 0000000000000286 00000001bf858770
18:30:54.466601500 kern.notice: Call Trace:
18:30:54.466621500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.466646500 kern.notice:  [<ffffffffa00ce712>] cv_wait_common+0x72/0xd0 [spl]
18:30:54.466668500 kern.notice:  [<ffffffff810b6090>] ? add_wait_queue+0x60/0x60
18:30:54.466690500 kern.notice:  [<ffffffffa00ce78e>] __cv_wait+0xe/0x10 [spl]
18:30:54.466714500 kern.notice:  [<ffffffffa0326983>] zio_wait+0xe3/0x150 [zfs]
18:30:54.466741500 kern.notice:  [<ffffffffa029fac7>] dbuf_read+0x337/0x740 [zfs]
18:30:54.466764500 kern.notice:  [<ffffffffa02a636b>] dmu_buf_hold+0xbb/0x170 [zfs]
18:30:54.466791500 kern.notice:  [<ffffffffa02fab5f>] zap_lockdir+0x4f/0x660 [zfs]
18:30:54.466814500 kern.notice:  [<ffffffffa02fc425>] zap_lookup_norm+0x45/0x190 [zfs]
18:30:54.466836500 kern.notice:  [<ffffffffa02fc5ee>] zap_lookup+0x2e/0x30 [zfs]
18:30:54.466858500 kern.notice:  [<ffffffffa0300f04>] zfs_dirent_lock+0x444/0x4f0 [zfs]
18:30:54.466880500 kern.notice:  [<ffffffffa0301018>] zfs_dirlook+0x68/0x210 [zfs]
18:30:54.466902500 kern.notice:  [<ffffffffa02ff6ce>] ? zfs_zaccess+0x9e/0x210 [zfs]
18:30:54.466928500 kern.notice:  [<ffffffffa0318ea1>] zfs_lookup+0x2f1/0x340 [zfs]
18:30:54.466950500 kern.notice:  [<ffffffffa032afd2>] zpl_lookup+0x52/0xb0 [zfs]
18:30:54.466971500 kern.notice:  [<ffffffff815e2999>] ? _raw_spin_lock+0x9/0x10
18:30:54.466993500 kern.notice:  [<ffffffff81186c39>] d_alloc_and_lookup+0x39/0x80
18:30:54.467015500 kern.notice:  [<ffffffff81188135>] do_lookup+0x2b5/0x3b0
18:30:54.467037500 kern.notice:  [<ffffffff812c3887>] ? security_inode_permission+0x17/0x20
18:30:54.467058500 kern.notice:  [<ffffffff81188ba7>] link_path_walk+0x137/0x860
18:30:54.467080500 kern.notice:  [<ffffffff81188995>] ? path_init+0x2e5/0x3c0
18:30:54.467102500 kern.notice:  [<ffffffff81189413>] path_lookupat+0x53/0x730
18:30:54.467123500 kern.notice:  [<ffffffff81189b1c>] do_path_lookup+0x2c/0xc0
18:30:54.467144500 kern.notice:  [<ffffffff8118b224>] user_path_at_empty+0x54/0xa0
18:30:54.467168500 kern.notice:  [<ffffffff81072f10>] ? do_page_fault+0x1f0/0x4a0
18:30:54.467190500 kern.notice:  [<ffffffff8118b27c>] user_path_at+0xc/0x10
18:30:54.467212500 kern.notice:  [<ffffffff81181185>] vfs_fstatat+0x35/0x60
18:30:54.467233500 kern.notice:  [<ffffffff8119acba>] ? mntput+0x1a/0x30
18:30:54.467254500 kern.notice:  [<ffffffff811811e6>] vfs_stat+0x16/0x20
18:30:54.467276500 kern.notice:  [<ffffffff810788e5>] sys32_stat64+0x15/0x30
18:30:54.467297500 kern.notice:  [<ffffffff815e3135>] ? page_fault+0x25/0x30
18:30:54.467319500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.467341500 kern.info: tcpsvd          D ffff880312035ee8     0 12791  16425 0x00020000
18:30:54.467362500 kern.notice:  ffff8803278d9ad8 0000000000000086 ffff880300000000 ffff8803278d9ef8
18:30:54.467387500 kern.notice:  ffff880312035b40 ffff8803278d9fd8 ffff8803278d9fd8 ffff8803278d9fd8
18:30:54.467410500 kern.notice:  ffff88040550ada0 ffff880312035b40 ffff8803278d9ab8 000000018020001b
18:30:54.467436500 kern.notice: Call Trace:
18:30:54.467456500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.467477500 kern.notice:  [<ffffffff815e1ac7>] __mutex_lock_slowpath+0xc7/0x140
18:30:54.467507500 kern.notice:  [<ffffffff81187a2e>] ? path_get+0x3e/0x50
18:30:54.467529500 kern.notice:  [<ffffffff815e16cd>] mutex_lock+0x1d/0x40
18:30:54.467551500 kern.notice:  [<ffffffff8118810b>] do_lookup+0x28b/0x3b0
18:30:54.467572500 kern.notice:  [<ffffffff812c3887>] ? security_inode_permission+0x17/0x20
18:30:54.467594500 kern.notice:  [<ffffffff81188ba7>] link_path_walk+0x137/0x860
18:30:54.467616500 kern.notice:  [<ffffffff81188995>] ? path_init+0x2e5/0x3c0
18:30:54.467638500 kern.notice:  [<ffffffff81189413>] path_lookupat+0x53/0x730
18:30:54.467659500 kern.notice:  [<ffffffff81189b1c>] do_path_lookup+0x2c/0xc0
18:30:54.467680500 kern.notice:  [<ffffffff8118b224>] user_path_at_empty+0x54/0xa0
18:30:54.467705500 kern.notice:  [<ffffffff81072f10>] ? do_page_fault+0x1f0/0x4a0
18:30:54.467727500 kern.notice:  [<ffffffff8118b27c>] user_path_at+0xc/0x10
18:30:54.467749500 kern.notice:  [<ffffffff81181185>] vfs_fstatat+0x35/0x60
18:30:54.467770500 kern.notice:  [<ffffffff8119acba>] ? mntput+0x1a/0x30
18:30:54.467792500 kern.notice:  [<ffffffff811811e6>] vfs_stat+0x16/0x20
18:30:54.467813500 kern.notice:  [<ffffffff810788e5>] sys32_stat64+0x15/0x30
18:30:54.467834500 kern.notice:  [<ffffffff815e3135>] ? page_fault+0x25/0x30
18:30:54.467856500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.467878500 kern.info: sleep           S ffff8801ae6bc818     0 14957   9295 0x00020000
18:30:54.467899500 kern.notice:  ffff8801b0e93e48 0000000000000082 0000000000000000 0000000100000000
18:30:54.467922500 kern.notice:  ffff8801ae6bc470 ffff8801b0e93fd8 ffff8801b0e93fd8 ffff8801b0e93fd8
18:30:54.467945500 kern.notice:  ffff8804054d5b40 ffff8801ae6bc470 ffffffff813030a0 0000000100000282
18:30:54.467970500 kern.notice: Call Trace:
18:30:54.467990500 kern.notice:  [<ffffffff813030a0>] ? rb_insert_color+0x100/0x140
18:30:54.468014500 kern.notice:  [<ffffffff815e0dca>] schedule+0x3a/0x50
18:30:54.468035500 kern.notice:  [<ffffffff815e1bd4>] do_nanosleep+0x94/0xd0
18:30:54.468057500 kern.notice:  [<ffffffff810ba4eb>] hrtimer_nanosleep+0x9b/0x140
18:30:54.468079500 kern.notice:  [<ffffffff810b9310>] ? update_rmtp+0x60/0x60
18:30:54.468100500 kern.notice:  [<ffffffff810df9ad>] compat_sys_nanosleep+0x8d/0x100
18:30:54.468122500 kern.notice:  [<ffffffff815e54d3>] ia32_do_call+0x13/0x13
18:30:54.468143500 kern.info: sleep           R  running task        0 14961   9291 0x00020000
18:30:54.468164500 kern.notice:  ffff8801b0ea3dd8 ffffffff810bc210 ffff88041fc73040 ffff88041fc73040
18:30:54.468187500 kern.notice:  ffff88041fc73040 0000000000000001 0000000000000001 000000000000006d
18:30:54.468210500 kern.notice:  ffff8801b0ea3e28 ffffffff8107e968 ffff880100000000 ffff8801b0ea3e28
18:30:54.468236500 kern.notice: Call Trace:
18:30:54.468256500 kern.notice:  [<ffffffff810bc210>] ? sched_clock_cpu+0xa0/0x100
18:30:54.468280500 kern.notice:  [<ffffffff8107e968>] ? enqueue_task+0x58/0x60
18:30:54.468302500 kern.notice:  [<ffffffff81090804>] ? sched_move_task+0x94/0x250
18:30:54.468324500 kern.notice:  [<ffffffff81052eea>] ? flush_ptrace_hw_breakpoint+0x1a/0x40
18:30:54.468346500 kern.notice:  [<ffffffff815e2c89>] ? _raw_write_lock_irq+0x19/0x20
18:30:54.468367500 kern.notice:  [<ffffffff81098344>] ? do_exit+0x1f4/0x8b0
18:30:54.468389500 kern.notice:  [<ffffffff81098d3f>] ? do_group_exit+0x3f/0xa0
18:30:54.468410500 kern.notice:  [<ffffffff81098db2>] ? sys_exit_group+0x12/0x20
18:30:54.468432500 kern.notice:  [<ffffffff815e54d3>] ? ia32_do_call+0x13/0x13
18:30:54.468453500 kern.info: sleep           R  running task        0 14965   9362 0x00000000
18:30:54.468478500 kern.notice:  0000000000000001 000000000000009b ffff8801b0ebde28 ffffffff8107e968
18:30:54.468501500 kern.notice:  ffff880100000000 ffff8801b0ebde28 0000000000000286 ffff8801ae6bdb40
18:30:54.468523500 kern.notice:  ffff8801b0ebde68 ffffffff81090804 ffff8803e425ad04 0000000000000286
18:30:54.468546500 kern.notice: Call Trace:
18:30:54.468566500 kern.notice:  [<ffffffff8107e968>] ? enqueue_task+0x58/0x60
18:30:54.468588500 kern.notice:  [<ffffffff81090804>] ? sched_move_task+0x94/0x250
18:30:54.468610500 kern.notice:  [<ffffffff81052eea>] ? flush_ptrace_hw_breakpoint+0x1a/0x40
18:30:54.468632500 kern.notice:  [<ffffffff815e2c89>] ? _raw_write_lock_irq+0x19/0x20
18:30:54.468654500 kern.notice:  [<ffffffff81098344>] ? do_exit+0x1f4/0x8b0
18:30:54.468678500 kern.notice:  [<ffffffff810ba4eb>] ? hrtimer_nanosleep+0x9b/0x140
18:30:54.468700500 kern.notice:  [<ffffffff81098d3f>] ? do_group_exit+0x3f/0xa0
18:30:54.468721500 kern.notice:  [<ffffffff81098db2>] ? sys_exit_group+0x12/0x20
18:30:54.468746500 kern.notice:  [<ffffffff815e35d2>] ? system_call_fastpath+0x16/0x1b
18:30:54.468768500 kern.notice: Sched Debug Version: v0.10, 3.1.6-vs2.3.2.5-newjayhawk #4
18:30:54.468790500 kern.notice: ktime                                   : 23612788.393336
18:30:54.468812500 kern.notice: sched_clk                               : 23564207.308220
18:30:54.468833500 kern.notice: cpu_clk                                 : 23536937.947522
18:30:54.468855500 kern.notice: jiffies                                 : 4297288751
18:30:54.468875500 kern.notice: sched_clock_stable                      : 0
18:30:54.468897500 kern.notice: 
18:30:54.468916500 kern.notice: sysctl_sched
18:30:54.468936500 kern.notice:   .sysctl_sched_latency                    : 18.000000
18:30:54.468957500 kern.notice:   .sysctl_sched_min_granularity            : 2.250000
18:30:54.468978500 kern.notice:   .sysctl_sched_wakeup_granularity         : 3.000000
18:30:54.469001500 kern.notice:   .sysctl_sched_child_runs_first           : 0
18:30:54.469026500 kern.notice:   .sysctl_sched_features                   : 15471
18:30:54.469048500 kern.notice:   .sysctl_sched_tunable_scaling            : 1 (logaritmic)
18:30:54.469098500 kern.notice: 
18:30:54.469117500 kern.notice: cpu#0, 2600.596 MHz
18:30:54.469137500 kern.notice:   .nr_running                    : 11
18:30:54.469158500 kern.notice:   .load                          : 89905
18:30:54.469181500 kern.notice:   .nr_switches                   : 104758198
18:30:54.469203500 kern.notice:   .nr_load_updates               : 2253478
18:30:54.469225500 kern.notice:   .nr_uninterruptible            : -105998
18:30:54.469246500 kern.notice:   .next_balance                  : 4297.288769
18:30:54.469267500 kern.notice:   .curr->pid                     : 5229
18:30:54.469288500 kern.notice:   .clock                         : 23101267.246679
18:30:54.469314500 kern.notice:   .cpu_load[0]                   : 89905
18:30:54.469336500 kern.notice:   .cpu_load[1]                   : 89905
18:30:54.469357500 kern.notice:   .cpu_load[2]                   : 89905
18:30:54.469381500 kern.notice:   .cpu_load[3]                   : 89905
18:30:54.469402500 kern.notice:   .cpu_load[4]                   : 89905
18:30:54.469424500 kern.notice:   .yld_count                     : 0
18:30:54.469445500 kern.notice:   .sched_switch                  : 0
18:30:54.469466500 kern.notice:   .sched_count                   : 105006895
18:30:54.469487500 kern.notice:   .sched_goidle                  : 47634911
18:30:54.469508500 kern.notice:   .avg_idle                      : 140188
18:30:54.469532500 kern.notice:   .ttwu_count                    : 223243736
18:30:54.469553500 kern.notice:   .ttwu_local                    : 37692607
18:30:54.469575500 kern.notice: 
18:30:54.469594500 kern.notice: cfs_rq[0]:/
18:30:54.469618500 kern.notice:   .exec_clock                    : 6801115.226499
18:30:54.469640500 kern.notice:   .MIN_vruntime                  : 54876070.877101
18:30:54.469661500 kern.notice:   .min_vruntime                  : 54876079.877101
18:30:54.469683500 kern.notice:   .max_vruntime                  : 54876101.117937
18:30:54.469704500 kern.notice:   .spread                        : 30.240836
18:30:54.469726500 kern.notice:   .spread0                       : 0.000000
18:30:54.469746500 kern.notice:   .nr_spread_over                : 71067
18:30:54.469767500 kern.notice:   .nr_running                    : 10
18:30:54.469788500 kern.notice:   .load                          : 89905
18:30:54.469810500 kern.notice:   .load_avg                      : 0.000000
18:30:54.469832500 kern.notice:   .load_period                   : 0.000000
18:30:54.469852500 kern.notice:   .load_contrib                  : 0
18:30:54.469876500 kern.notice:   .load_tg                       : 0
18:30:54.469901500 kern.notice: 
18:30:54.469921500 kern.notice: rt_rq[0]:/irc
18:30:54.469940500 kern.notice:   .rt_nr_running                 : 0
18:30:54.469960500 kern.notice:   .rt_throttled                  : 0
18:30:54.469982500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470005500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470026500 kern.notice: 
18:30:54.470046500 kern.notice: rt_rq[0]:/stop32
18:30:54.470066500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470087500 kern.notice:   .rt_throttled                  : 0
18:30:54.470109500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470131500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470153500 kern.notice: 
18:30:54.470172500 kern.notice: rt_rq[0]:/webfrontend
18:30:54.470193500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470214500 kern.notice:   .rt_throttled                  : 0
18:30:54.470238500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470259500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470281500 kern.notice: 
18:30:54.470300500 kern.notice: rt_rq[0]:/tracsvn
18:30:54.470320500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470342500 kern.notice:   .rt_throttled                  : 0
18:30:54.470366500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470387500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470409500 kern.notice: 
18:30:54.470428500 kern.notice: rt_rq[0]:/postgres-bud
18:30:54.470449500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470470500 kern.notice:   .rt_throttled                  : 0
18:30:54.470492500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470513500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470537500 kern.notice: 
18:30:54.470556500 kern.notice: rt_rq[0]:/nethack
18:30:54.470576500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470597500 kern.notice:   .rt_throttled                  : 0
18:30:54.470620500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470642500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470663500 kern.notice: 
18:30:54.470682500 kern.notice: rt_rq[0]:/autobuild64
18:30:54.470702500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470723500 kern.notice:   .rt_throttled                  : 0
18:30:54.470748500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470770500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470791500 kern.notice: 
18:30:54.470810500 kern.notice: rt_rq[0]:/autobuild32
18:30:54.470830500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470851500 kern.notice:   .rt_throttled                  : 0
18:30:54.470876500 kern.notice:   .rt_time                       : 0.000000
18:30:54.470898500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.470919500 kern.notice: 
18:30:54.470938500 kern.notice: rt_rq[0]:/ldap-stop
18:30:54.470959500 kern.notice:   .rt_nr_running                 : 0
18:30:54.470980500 kern.notice:   .rt_throttled                  : 0
18:30:54.471001500 kern.notice:   .rt_time                       : 0.000000
18:30:54.471023500 kern.notice:   .rt_runtime                    : 0.000000
18:30:54.471045500 kern.notice: 
18:30:54.471064500 kern.notice: rt_rq[0]:/
18:30:54.471083500 kern.notice:   .rt_nr_running                 : 1
18:30:54.471105500 kern.notice:   .rt_throttled                  : 0
18:30:54.471126500 kern.notice:   .rt_time                       : 0.000000
18:30:54.471147500 kern.notice:   .rt_runtime                    : 1000.000000
18:30:54.471169500 kern.notice: 
18:30:54.471191500 kern.notice: runnable tasks:
18:30:54.471213500 kern.notice:             task   PID         tree-key  switches  prio     exec-runtime         sum-exec        sum-sleep
18:30:54.471235500 kern.notice: ----------------------------------------------------------------------------------------------------------
18:30:54.471261500 kern.notice:           zvol/0  1974  54876070.877101     11505   139  54876070.877101      1925.199361  22801532.099817 /
18:30:54.471286500 kern.notice:           zvol/6  1980  54876070.877101     11298   139  54876070.877101      1352.794378  22819698.098895 /
18:30:54.471311500 kern.notice:          zvol/12  1986  54876070.877101     11570   139  54876070.877101      1385.186251  22803354.053110 /
18:30:54.471336500 kern.notice:     z_null_iss/0  2156  54876073.234144     17985   139  54876073.234144      1423.381919  22240042.920208 /
18:30:54.471360500 kern.notice:     z_null_iss/0  5059  54876080.049547    693716   139  54876080.049547     35685.764170  21617470.967462 /
18:30:54.471386500 kern.notice:     z_null_int/0  5060  54876101.117937    205363   139  54876101.117937     27917.252154  22376006.905359 /
18:30:54.471412500 kern.notice:       z_rd_int/0  5069  54876076.457346  24551804   139  54876076.457346    560092.804797  19858520.120036 /
18:30:54.471436500 kern.notice:      z_wr_int/12  5098  54876079.877101   2493410   139  54876079.877101    128083.341139  21647073.239116 /
18:30:54.471461500 kern.notice: Rzfs_iput_taskq/  5229  54876078.243965     67152   100  54876078.243965    140384.379792  22823016.134943 /
18:30:54.471487500 kern.notice:         watchdog  5580         0.000000     43549    98         0.000000   1195290.608874         0.021196 /
18:30:54.471513500 kern.notice:      kworker/0:1  6151  54876072.373688       622   120  54876072.373688       138.543285    238807.749058 /
18:30:54.471538500 kern.notice: 
18:30:54.471557500 kern.notice: cpu#1, 2600.596 MHz
18:30:54.471580500 kern.notice:   .nr_running                    : 1
18:30:54.471602500 kern.notice:   .load                          : 1024
18:30:54.471623500 kern.notice:   .nr_switches                   : 82465614
18:30:54.471644500 kern.notice:   .nr_load_updates               : 2169600
18:30:54.471665500 kern.notice:   .nr_uninterruptible            : -54792
18:30:54.471686500 kern.notice:   .next_balance                  : 4297.288773
18:30:54.471708500 kern.notice:   .curr->pid                     : 16425
18:30:54.471729500 kern.notice:   .clock                         : 23582641.345224
18:30:54.471751500 kern.notice:   .cpu_load[0]                   : 0
18:30:54.471772500 kern.notice:   .cpu_load[1]                   : 0
18:30:54.471793500 kern.notice:   .cpu_load[2]                   : 3
18:30:54.471813500 kern.notice:   .cpu_load[3]                   : 49
18:30:54.471834500 kern.notice:   .cpu_load[4]                   : 109
18:30:54.471862500 kern.notice:   .yld_count                     : 0
18:30:54.471883500 kern.notice:   .sched_switch                  : 0
18:30:54.471904500 kern.notice:   .sched_count                   : 82567172
18:30:54.471925500 kern.notice:   .sched_goidle                  : 36328067
18:30:54.471946500 kern.notice:   .avg_idle                      : 1000000
18:30:54.471967500 kern.notice:   .ttwu_count                    : 12670219
18:30:54.471989500 kern.notice:   .ttwu_local                    : 2928251
18:30:54.472010500 kern.notice: 
18:30:54.472030500 kern.notice: cfs_rq[1]:/stop32
18:30:54.472050500 kern.notice:   .exec_clock                    : 547286.203664
18:30:54.472070500 kern.notice:   .MIN_vruntime                  : 0.000001
18:30:54.472092500 kern.notice:   .min_vruntime                  : 586504.730672
18:30:54.472114500 kern.notice:   .max_vruntime                  : 0.000001
18:30:54.472134500 kern.notice:   .spread                        : 0.000000
18:30:54.472158500 kern.notice:   .spread0                       : -54289575.146429
18:30:54.472179500 kern.notice:   .nr_spread_over                : 24565
18:30:54.472200500 kern.notice:   .nr_running                    : 1
18:30:54.472221500 kern.notice:   .load                          : 1024
18:30:54.472242500 kern.notice:   .load_avg                      : 0.000000
18:30:54.472263500 kern.notice:   .load_period                   : 9.999999
18:30:54.472288500 kern.notice:   .load_contrib                  : 0
18:30:54.472310500 kern.notice:   .load_tg                       : 0
18:30:54.472330500 kern.notice:   .se->exec_start                : 23582641.265140
18:30:54.472351500 kern.notice:   .se->vruntime                  : 54267796.370972
18:30:54.472372500 kern.notice:   .se->sum_exec_runtime          : 547295.849707
18:30:54.472394500 kern.notice:   .se->statistics.wait_start     : 0.000000
18:30:54.472416500 kern.notice:   .se->statistics.sleep_start    : 0.000000
18:30:54.472441500 kern.notice:   .se->statistics.block_start    : 0.000000
18:30:54.472462500 kern.notice:   .se->statistics.sleep_max      : 0.000000
18:30:54.472483500 kern.notice:   .se->statistics.block_max      : 0.000000
18:30:54.472504500 kern.notice:   .se->statistics.exec_max       : 66.315469
18:30:54.472528500 kern.notice:   .se->statistics.slice_max      : 53.472903
18:30:54.472550500 kern.notice:   .se->statistics.wait_max       : 209.414696
18:30:54.472571500 kern.notice:   .se->statistics.wait_sum       : 6791.047936
18:30:54.472592500 kern.notice:   .se->statistics.wait_count     : 281862
18:30:54.472613500 kern.notice:   .se->load.weight               : 1024
18:30:54.472634500 kern.notice: 
18:30:54.472653500 kern.notice: cfs_rq[1]:/autogroup-11
18:30:54.472674500 kern.notice:   .exec_clock                    : 829.918059
18:30:54.472696500 kern.notice:   .MIN_vruntime                  : 0.000001
18:30:54.472720500 kern.notice:   .min_vruntime                  : 828.869483
18:30:54.472743500 kern.notice:   .max_vruntime                  : 0.000001
18:30:54.472764500 kern.notice:   .spread                        : 0.000000
18:30:54.472785500 kern.notice:   .spread0                       : -54875251.007618
18:30:54.472805500 kern.notice:   .nr_spread_over                : 0
18:30:54.472826500 kern.notice:   .nr_running                    : 0
18:30:54.472847500 kern.notice:   .load                          : 0
18:30:54.472868500 kern.notice:   .load_avg                      : 5119.999488
18:30:54.472890500 kern.notice:   .load_period                   : 5.001505
18:30:54.472911500 kern.notice:   .load_contrib                  : 1023
18:30:54.472933500 kern.notice:   .load_tg                       : 1023
18:30:54.472954500 kern.notice:   .se->exec_start                : 23582314.617194
18:30:54.472976500 kern.notice:   .se->vruntime                  : 54267796.513287
18:30:54.473001500 kern.notice:   .se->sum_exec_runtime          : 830.167531
18:30:54.473022500 kern.notice:   .se->statistics.wait_start     : 0.000000
18:30:54.473043500 kern.notice:   .se->statistics.sleep_start    : 0.000000
18:30:54.473068500 kern.notice:   .se->statistics.block_start    : 0.000000
18:30:54.473090500 kern.notice:   .se->statistics.sleep_max      : 0.000000
18:30:54.473111500 kern.notice:   .se->statistics.block_max      : 0.000000
18:30:54.473134500 kern.notice:   .se->statistics.exec_max       : 14.045379
18:30:54.473156500 kern.notice:   .se->statistics.slice_max      : 0.426809
18:30:54.473176500 kern.notice:   .se->statistics.wait_max       : 32.414572
18:30:54.473198500 kern.noti

@akorn
Copy link
Contributor Author

akorn commented Jan 8, 2012

Allow me to quantify "it's taking forever". This is from strace -tttT on the rsync:

1326053450.822179 open("nstxd/.svn/prop-base/.finish.svn-base.Bf3YZo", O_RDWR|O_CREAT|O_EXCL, 0600) = 1 <0.068796>
1326053450.891107 fchmod(1, 0600)       = 0 <140.069799>
1326053590.961229 write(1, "K 14\nsvn:executable\nV 1\n*\nEND\n", 30) = 30 <0.000130>
1326053590.961437 close(1)              = 0 <0.000019>

That's right: the fchmod() took more than two minutes to complete. It's not always fchmod(); sometimes it's rename(), other times chmod() or even open() that takes so long.

@akorn
Copy link
Contributor Author

akorn commented Jan 8, 2012

Load levels are back to normal once zfs destroy completed (the first time around it said "dataset is busy" after 3,5 hours even though it wasn't busy, but it completed relatively quickly and successfully the second time around).

rsync speed is back up to normal-ish as well.

@akorn
Copy link
Contributor Author

akorn commented Jan 8, 2012

The same cron.weekly experiment results in negligible load and certainly no deadlock with the rootfs of the vservers residing on ext4 over LVM instead of ext4 over zvol, even though I added in all daily cronjobs as well to stress the system further. (Granted, it's on an SSD now whereas it used to be on raidz2 before, which could explain up to two orders of magnitude difference in performance, but what I'm seeing here is a lot more than that -- plus, no deadlock.)

@behlendorf
Copy link
Contributor

The original issue has has been resolved by commit 0d5c500 which will be part of 0.6.4.

@behlendorf behlendorf modified the milestones: 0.6.4, 0.7.0 Oct 6, 2014
ryao added a commit to ryao/zfs that referenced this issue Oct 9, 2014
The below excerpt of a backtrace is from a ztest failure when running
ZoL's ztest.

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

The backtrace was an infinite loop of `vdev_queue_io_to_issue()` invoking
`zio_execute()` until it overran the stack. vdev_queue_io_to_issue() will ony
invoke `zio_execute()` on raidz vdevs when aggregation I/Os are generated to
improve aggregation continuity. These I/Os do not trigger any writes. However,
it appears that they can be generated in such a way that they recurse
infinitely upon return to `vdev_queue_io_to_issue()`. As a consequence, we see
the number of parents by 1 each time the recursion returns to
`vdev_raidz_io_start()`.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 9, 2014
The below excerpt of a backtrace is from a ztest failure when running
ZoL's ztest.

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

The backtrace was an infinite loop of `vdev_queue_io_to_issue()` invoking
`zio_execute()` until it overran the stack. vdev_queue_io_to_issue() will ony
invoke `zio_execute()` on raidz vdevs when aggregation I/Os are generated to
improve aggregation continuity. These I/Os do not trigger any writes. However,
it appears that they can be generated in such a way that they recurse
infinitely upon return to `vdev_queue_io_to_issue()`. As a consequence, we see
the number of parents by 1 each time the recursion returns to
`vdev_raidz_io_start()`.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 9, 2014
The below excerpt of a backtrace is from a ztest failure when running
ZoL's ztest.

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

The backtrace was an infinite loop of `vdev_queue_io_to_issue()` invoking
`zio_execute()` until it overran the stack. vdev_queue_io_to_issue() will ony
invoke `zio_execute()` on raidz vdevs when aggregation I/Os are generated to
improve aggregation continuity. These I/Os do not trigger any writes. However,
it appears that they can be generated in such a way that they recurse
infinitely upon return to `vdev_queue_io_to_issue()`.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 10, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 10, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 10, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 10, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 10, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 10, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 11, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
ryao added a commit to ryao/zfs that referenced this issue Oct 11, 2014
The below excerpt of a backtrace is from a ztest failure when running ZoL's
ztest:

/openzfs#453 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#454 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350003de0) at ../../module/zfs/vdev_queue.c:747
/openzfs#455 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350003de0) at ../../module/zfs/zio.c:2659
/openzfs#456 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1399
/openzfs#457 zio_nowait (zio=0x7f0350003de0) at ../../module/zfs/zio.c:1456
/openzfs#458 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350003a10) at ../../module/zfs/vdev_mirror.c:374
/openzfs#459 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1399
/openzfs#460 zio_nowait (zio=0x7f0350003a10) at ../../module/zfs/zio.c:1456
/openzfs#461 0x00007f03c806464c in vdev_raidz_io_start (zio=0x7f0350003380) at ../../module/zfs/vdev_raidz.c:1607
/openzfs#462 0x00007f03c807f243 in __zio_execute (zio=0x7f0350003380) at ../../module/zfs/zio.c:1399
/openzfs#463 zio_nowait (zio=0x7f0350003380) at ../../module/zfs/zio.c:1456
/openzfs#464 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002fb0) at ../../module/zfs/vdev_mirror.c:374
/openzfs#465 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1399
/openzfs#466 zio_nowait (zio=0x7f0350002fb0) at ../../module/zfs/zio.c:1456
/openzfs#467 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033957ebf0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#468 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:2707
/openzfs#469 0x00007f03c808285b in __zio_execute (zio=0x7f033957ebf0) at ../../module/zfs/zio.c:1399
/openzfs#470 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f0390001330, pio=0x7f033957ebf0) at ../../module/zfs/zio.c:547
/openzfs#471 zio_done (zio=0x7f0390001330) at ../../module/zfs/zio.c:3278
/openzfs#472 0x00007f03c808285b in __zio_execute (zio=0x7f0390001330) at ../../module/zfs/zio.c:1399
/openzfs#473 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4013a00, pio=0x7f0390001330) at ../../module/zfs/zio.c:547
/openzfs#474 zio_done (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:3278
/openzfs#475 0x00007f03c808285b in __zio_execute (zio=0x7f03b4013a00) at ../../module/zfs/zio.c:1399
/openzfs#476 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014210, pio=0x7f03b4013a00) at ../../module/zfs/zio.c:547
/openzfs#477 zio_done (zio=0x7f03b4014210) at ../../module/zfs/zio.c:3278
/openzfs#478 0x00007f03c808285b in __zio_execute (zio=0x7f03b4014210) at ../../module/zfs/zio.c:1399
/openzfs#479 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03b4014620, pio=0x7f03b4014210) at ../../module/zfs/zio.c:547
/openzfs#480 zio_done (zio=0x7f03b4014620) at ../../module/zfs/zio.c:3278
/openzfs#481 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03b4014620) at ../../module/zfs/zio.c:1399
/openzfs#482 zio_execute (zio=zio@entry=0x7f03b4014620) at ../../module/zfs/zio.c:1337
/openzfs#483 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#484 0x00007f03c806106e in vdev_queue_io (zio=zio@entry=0x7f0350002be0) at ../../module/zfs/vdev_queue.c:747
/openzfs#485 0x00007f03c80818c1 in zio_vdev_io_start (zio=0x7f0350002be0) at ../../module/zfs/zio.c:2659
/openzfs#486 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1399
/openzfs#487 zio_nowait (zio=0x7f0350002be0) at ../../module/zfs/zio.c:1456
/openzfs#488 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350002810) at ../../module/zfs/vdev_mirror.c:374
/openzfs#489 0x00007f03c807f243 in __zio_execute (zio=0x7f0350002810) at ../../module/zfs/zio.c:1399
/openzfs#490 zio_nowait (zio=0x7f0350002810) at ../../module/zfs/zio.c:1456
/openzfs#491 0x00007f03c8064593 in vdev_raidz_io_start (zio=0x7f0350001270) at ../../module/zfs/vdev_raidz.c:1591
/openzfs#492 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001270) at ../../module/zfs/zio.c:1399
/openzfs#493 zio_nowait (zio=0x7f0350001270) at ../../module/zfs/zio.c:1456
/openzfs#494 0x00007f03c805f71b in vdev_mirror_io_start (zio=0x7f0350001e60) at ../../module/zfs/vdev_mirror.c:374
/openzfs#495 0x00007f03c807f243 in __zio_execute (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1399
/openzfs#496 zio_nowait (zio=0x7f0350001e60) at ../../module/zfs/zio.c:1456
/openzfs#497 0x00007f03c805ed43 in vdev_mirror_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/vdev_mirror.c:499
/openzfs#498 0x00007f03c807a0c0 in zio_vdev_io_done (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:2707
/openzfs#499 0x00007f03c808285b in __zio_execute (zio=0x7f033a0c39c0) at ../../module/zfs/zio.c:1399
/openzfs#500 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f03a8003c00, pio=0x7f033a0c39c0) at ../../module/zfs/zio.c:547
/openzfs#501 zio_done (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:3278
/openzfs#502 0x00007f03c808285b in __zio_execute (zio=0x7f03a8003c00) at ../../module/zfs/zio.c:1399
/openzfs#503 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800c400, pio=0x7f03a8003c00) at ../../module/zfs/zio.c:547
/openzfs#504 zio_done (zio=0x7f038800c400) at ../../module/zfs/zio.c:3278
/openzfs#505 0x00007f03c808285b in __zio_execute (zio=0x7f038800c400) at ../../module/zfs/zio.c:1399
/openzfs#506 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800da00, pio=0x7f038800c400) at ../../module/zfs/zio.c:547
/openzfs#507 zio_done (zio=0x7f038800da00) at ../../module/zfs/zio.c:3278
/openzfs#508 0x00007f03c808285b in __zio_execute (zio=0x7f038800da00) at ../../module/zfs/zio.c:1399
/openzfs#509 zio_notify_parent (wait=ZIO_WAIT_DONE, zio=0x7f038800fd80, pio=0x7f038800da00) at ../../module/zfs/zio.c:547
/openzfs#510 zio_done (zio=0x7f038800fd80) at ../../module/zfs/zio.c:3278
/openzfs#511 0x00007f03c807a6d3 in __zio_execute (zio=0x7f038800fd80) at ../../module/zfs/zio.c:1399
/openzfs#512 zio_execute (zio=zio@entry=0x7f038800fd80) at ../../module/zfs/zio.c:1337
/openzfs#513 0x00007f03c8060b35 in vdev_queue_io_to_issue (vq=vq@entry=0x99f8a8) at ../../module/zfs/vdev_queue.c:706
/openzfs#514 0x00007f03c806119d in vdev_queue_io_done (zio=zio@entry=0x7f03a0010950) at ../../module/zfs/vdev_queue.c:775
/openzfs#515 0x00007f03c807a0e8 in zio_vdev_io_done (zio=0x7f03a0010950) at ../../module/zfs/zio.c:2686
/openzfs#516 0x00007f03c807a6d3 in __zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1399
/openzfs#517 zio_execute (zio=0x7f03a0010950) at ../../module/zfs/zio.c:1337
/openzfs#518 0x00007f03c7fcd0c4 in taskq_thread (arg=0x966d50) at ../../lib/libzpool/taskq.c:215
/openzfs#519 0x00007f03c7fc7937 in zk_thread_helper (arg=0x967e90) at ../../lib/libzpool/kernel.c:135
/openzfs#520 0x00007f03c78890a3 in start_thread (arg=0x7f03c2703700) at pthread_create.c:309
/openzfs#521 0x00007f03c75c50fd in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111

This occurred when ztest was simulating a scrub under heavy I/O load. Under
those circumstances, it was possible for a mix of noop I/Os for aggregation
continuity and the I/O elevator to generate arbitrarily deep recursion.

This patch modifies ZFS to propapage a recursion counter inside the zio_t
objects such that IOs will be redispatched upon reaching a given recursion
depth.  We can detect long call chains and dispatch to another ZIO taskq. We
cut in-line when we do this to minimize the potential for taskq exhaustion that
can prevent a zio from notifying its parent.

Signed-off-by: Richard Yao <[email protected]>
sdimitro pushed a commit to sdimitro/zfs that referenced this issue Jul 18, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants