Re: Possible regression with cgroups in 3.11

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I am sorry, but kdump crash files are difficult to obtain on our
systems, since we are using nfsroot on diskless clients. Is there any
possibility to see, why "synchronize_rcu" is actually waiting? I tried
looking through the code but did not get very far. In any case, I am
appending current stack dumps from kernel 3.11.6. With lockdep
enabled, there were also no additional warnings in the kernel log.

The thread with "mem_cgroup_reparent_charges" is hanging at synchronize_rcu:

crash> bt -t 1200
PID: 1200   TASK: ffff883ff9db9770  CPU: 56  COMMAND: "kworker/56:0"
              START: __schedule at ffffffff813bb12c
  [ffff883ef84ffbd8] schedule at ffffffff813bb2cc
  [ffff883ef84ffbe8] schedule_timeout at ffffffff813b9234
  [ffff883ef84ffbf8] __wake_up_common at ffffffff8104a8bd
  [ffff883ef84ffc30] _raw_spin_unlock_irqrestore at ffffffff813bc55b
  [ffff883ef84ffc60] __wait_for_common at ffffffff813bab7f
  [ffff883ef84ffc68] schedule_timeout at ffffffff813b9200
  [ffff883ef84ffc80] default_wake_function at ffffffff8104eec3
  [ffff883ef84ffc98] call_rcu at ffffffff810937ff
  [ffff883ef84ffcc8] wait_for_completion at ffffffff813bac1b
  [ffff883ef84ffcd8] wait_rcu_gp at ffffffff81041ea6
  [ffff883ef84ffce8] wakeme_after_rcu at ffffffff81041e51
  [ffff883ef84ffd20] synchronize_rcu at ffffffff81092333
  [ffff883ef84ffd30] mem_cgroup_reparent_charges at ffffffff810e3962
  [ffff883ef84ffdc0] mem_cgroup_css_offline at ffffffff810e3d6e
  [ffff883ef84ffdf0] offline_css at ffffffff8107a872
  [ffff883ef84ffe10] cgroup_offline_fn at ffffffff8107c55f
  [ffff883ef84ffe50] process_one_work at ffffffff8103f26f
  [ffff883ef84ffe90] worker_thread at ffffffff8103f711
  [ffff883ef84ffeb0] worker_thread at ffffffff8103f5cd
  [ffff883ef84ffec8] kthread at ffffffff810441a4
  [ffff883ef84fff28] kthread at ffffffff8104411c
  [ffff883ef84fff50] ret_from_fork at ffffffff813bd02c
  [ffff883ef84fff80] kthread at ffffffff8104411c

The other stack traces from waiting threads are identical to these:

crash> bt -t 6721
PID: 6721   TASK: ffff8834940b5dc0  CPU: 11  COMMAND: "lssubsys"
              START: __schedule at ffffffff813bb12c
  [ffff8831e01d5dc8] schedule at ffffffff813bb2cc
  [ffff8831e01d5dd8] schedule_preempt_disabled at ffffffff813bb553
  [ffff8831e01d5de8] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff8831e01d5e40] mutex_lock at ffffffff813b9640
  [ffff8831e01d5e58] proc_cgroupstats_show at ffffffff8107a0d7
  [ffff8831e01d5e78] seq_read at ffffffff8110492b
  [ffff8831e01d5ea8] acct_account_cputime at ffffffff81096a99
  [ffff8831e01d5ee0] proc_reg_read at ffffffff811325e0
  [ffff8831e01d5f18] vfs_read at ffffffff810eaaa3
  [ffff8831e01d5f48] sys_read at ffffffff810eb18f
  [ffff8831e01d5f80] tracesys at ffffffff813bd2cb
    RIP: 00007ffe7cdd1c50  RSP: 00007fffe43ec9b8  RFLAGS: 00000246
    RAX: ffffffffffffffda  RBX: ffffffff813bd2cb  RCX: ffffffffffffffff
    RDX: 0000000000000400  RSI: 00007ffe7d730000  RDI: 0000000000000002
    RBP: 000000000114a250   R8: 00000000ffffffff   R9: 0000000000000000
    R10: 0000000000000022  R11: 0000000000000246  R12: 0000000000000000
    R13: 000000000000000a  R14: 000000000114a010  R15: 0000000000000000
    ORIG_RAX: 0000000000000000  CS: 0033  SS: 002b
crash> bt -t 6618
PID: 6618   TASK: ffff8807e645ddc0  CPU: 5   COMMAND: "kworker/5:1"
              START: __schedule at ffffffff813bb12c
  [ffff880396c4fd98] schedule at ffffffff813bb2cc
  [ffff880396c4fda8] schedule_preempt_disabled at ffffffff813bb553
  [ffff880396c4fdb8] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff880396c4fdd8] mmdrop at ffffffff8104b2ce
  [ffff880396c4fe10] mutex_lock at ffffffff813b9640
  [ffff880396c4fe28] cgroup_free_fn at ffffffff81079e3e
  [ffff880396c4fe50] process_one_work at ffffffff8103f26f
  [ffff880396c4fe90] worker_thread at ffffffff8103f711
  [ffff880396c4feb0] worker_thread at ffffffff8103f5cd
  [ffff880396c4fec8] kthread at ffffffff810441a4
  [ffff880396c4ff28] kthread at ffffffff8104411c
  [ffff880396c4ff50] ret_from_fork at ffffffff813bd02c
  [ffff880396c4ff80] kthread at ffffffff8104411c
crash> bt -t 3053
PID: 3053   TASK: ffff881ffb724650  CPU: 50  COMMAND: "slurmstepd"
              START: __schedule at ffffffff813bb12c
  [ffff881e2e7b7dc8] schedule at ffffffff813bb2cc
  [ffff881e2e7b7dd8] schedule_preempt_disabled at ffffffff813bb553
  [ffff881e2e7b7de8] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff881e2e7b7e08] shrink_dcache_parent at ffffffff810faf35
  [ffff881e2e7b7e40] mutex_lock at ffffffff813b9640
  [ffff881e2e7b7e58] cgroup_rmdir at ffffffff8107d42b
  [ffff881e2e7b7e78] vfs_rmdir at ffffffff810f5dea
  [ffff881e2e7b7ea0] do_rmdir at ffffffff810f5eff
  [ffff881e2e7b7f28] syscall_trace_enter at ffffffff8100c195
  [ffff881e2e7b7f70] sys_rmdir at ffffffff810f6c42
  [ffff881e2e7b7f80] tracesys at ffffffff813bd2cb
    RIP: 00007fa31ca8c047  RSP: 00007fffaa493f08  RFLAGS: 00000202
    RAX: ffffffffffffffda  RBX: ffffffff813bd2cb  RCX: ffffffffffffffff
    RDX: 0000000000000000  RSI: 0000000000000002  RDI: 000000000133b408
    RBP: 0000000000000000   R8: 0000000000000019   R9: 0101010101010101
    R10: 00007fffaa493ce0  R11: 0000000000000202  R12: ffffffff810f6c42
    R13: ffff881e2e7b7f78  R14: 00007fffaa494000  R15: 000000000132f758
    ORIG_RAX: 0000000000000054  CS: 0033  SS: 002b
crash> bt -t 1224
PID: 1224   TASK: ffff8807e646aee0  CPU: 7   COMMAND: "kworker/7:0"
              START: __schedule at ffffffff813bb12c
  [ffff88010cd6fd78] schedule at ffffffff813bb2cc
  [ffff88010cd6fd88] schedule_preempt_disabled at ffffffff813bb553
  [ffff88010cd6fd98] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff88010cd6fdb8] _raw_spin_unlock_irqrestore at ffffffff813bc55b
  [ffff88010cd6fdf8] mutex_lock at ffffffff813b9640
  [ffff88010cd6fe10] cgroup_offline_fn at ffffffff8107c523
  [ffff88010cd6fe50] process_one_work at ffffffff8103f26f
  [ffff88010cd6fe90] worker_thread at ffffffff8103f711
  [ffff88010cd6feb0] worker_thread at ffffffff8103f5cd
  [ffff88010cd6fec8] kthread at ffffffff810441a4
  [ffff88010cd6ff28] kthread at ffffffff8104411c
  [ffff88010cd6ff50] ret_from_fork at ffffffff813bd02c
  [ffff88010cd6ff80] kthread at ffffffff8104411c
crash> bt -t 1159
PID: 1159   TASK: ffff8807e5455dc0  CPU: 5   COMMAND: "kworker/5:0"
              START: __schedule at ffffffff813bb12c
  [ffff88031fdefd98] schedule at ffffffff813bb2cc
  [ffff88031fdefda8] schedule_preempt_disabled at ffffffff813bb553
  [ffff88031fdefdb8] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff88031fdefdd8] mmdrop at ffffffff8104b2ce
  [ffff88031fdefe10] mutex_lock at ffffffff813b9640
  [ffff88031fdefe28] cgroup_free_fn at ffffffff81079e3e
  [ffff88031fdefe50] process_one_work at ffffffff8103f26f
  [ffff88031fdefe90] worker_thread at ffffffff8103f711
  [ffff88031fdefeb0] worker_thread at ffffffff8103f5cd
  [ffff88031fdefec8] kthread at ffffffff810441a4
  [ffff88031fdeff28] kthread at ffffffff8104411c
  [ffff88031fdeff50] ret_from_fork at ffffffff813bd02c
  [ffff88031fdeff80] kthread at ffffffff8104411c
crash> bt -t 2449
PID: 2449   TASK: ffff881ffb0aaee0  CPU: 31  COMMAND: "kworker/31:1"
              START: __schedule at ffffffff813bb12c
  [ffff881ffad2dd68] schedule at ffffffff813bb2cc
  [ffff881ffad2dd78] schedule_preempt_disabled at ffffffff813bb553
  [ffff881ffad2dd88] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff881ffad2dde0] mutex_lock at ffffffff813b9640
  [ffff881ffad2ddf8] cgroup_release_agent at ffffffff8107b8a1
  [ffff881ffad2de50] process_one_work at ffffffff8103f26f
  [ffff881ffad2de90] worker_thread at ffffffff8103f711
  [ffff881ffad2deb0] worker_thread at ffffffff8103f5cd
  [ffff881ffad2dec8] kthread at ffffffff810441a4
  [ffff881ffad2df28] kthread at ffffffff8104411c
  [ffff881ffad2df50] ret_from_fork at ffffffff813bd02c
  [ffff881ffad2df80] kthread at ffffffff8104411c
crash> bt -t 1130
PID: 1130   TASK: ffff8827fb051770  CPU: 35  COMMAND: "kworker/35:1"
              START: __schedule at ffffffff813bb12c
  [ffff8827fb7d9d98] schedule at ffffffff813bb2cc
  [ffff8827fb7d9da8] schedule_preempt_disabled at ffffffff813bb553
  [ffff8827fb7d9db8] __mutex_lock_slowpath at ffffffff813ba46f
  [ffff8827fb7d9dd8] mmdrop at ffffffff8104b2ce
  [ffff8827fb7d9e10] mutex_lock at ffffffff813b9640
  [ffff8827fb7d9e28] cgroup_free_fn at ffffffff81079e3e
  [ffff8827fb7d9e50] process_one_work at ffffffff8103f26f
  [ffff8827fb7d9e90] worker_thread at ffffffff8103f711
  [ffff8827fb7d9eb0] worker_thread at ffffffff8103f5cd
  [ffff8827fb7d9ec8] kthread at ffffffff810441a4
  [ffff8827fb7d9f28] kthread at ffffffff8104411c
  [ffff8827fb7d9f50] ret_from_fork at ffffffff813bd02c
  [ffff8827fb7d9f80] kthread at ffffffff8104411c
--
To unsubscribe from this list: send the line "unsubscribe cgroups" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]     [Monitors]

  Powered by Linux