[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Bug#602273: linux-image-2.6.32-5-686-bigmem: domU hangs during dom0 reboot, recovers when dom0 uptime "caught up"



Hi,

I seem to have hit the same bug again with 2.6.32-38 domU (amd64). After
I rebooted dom0 to 2.6.32-41 I see that the jiffies value of the domU
does not increase but cpu_time in xm list -l output does. Value of
"last_value" seems to increase (now around 1804512135271).

I looked at linux-source-2.6.32 2.6.32-38 and I see that it contains a
call to pvclock_resume in xen_timer_resume so the patch should still be
there. Is this a new bug that just has the same symptoms?

      KERNEL: /usr/lib/debug/boot/vmlinux-2.6.32-5-amd64
    DUMPFILE: /local/xen/lindi1/core
        CPUS: 6
        DATE: Mon Mar  5 10:02:26 2012
      UPTIME: 54 days, 21:35:49
LOAD AVERAGE: 0.04, 0.01, 0.00
       TASKS: 307
    NODENAME: lindi1
     RELEASE: 2.6.32-5-amd64
     VERSION: #1 SMP Mon Oct 3 03:59:20 UTC 2011
     MACHINE: x86_64  (3210 Mhz)
      MEMORY: 2 GB
       PANIC: ""
         PID: 0
     COMMAND: "swapper"
        TASK: ffffffff814611f0  (1 of 6)  [THREAD_INFO: ffffffff8142c000]
         CPU: 0
       STATE: TASK_RUNNING (ACTIVE)
     WARNING: panic task not found


crash> disassemble xen_timer_resume
Dump of assembler code for function xen_timer_resume:
0xffffffff8100de32 <xen_timer_resume+0>:        push   %rbx
0xffffffff8100de33 <xen_timer_resume+1>:        callq  0xffffffff8102cd7e <pvclock_resume>
                                                                           ^^^^^^^^^^^^^^
...

crash> disassemble pvclock_resume
Dump of assembler code for function pvclock_resume:
0xffffffff8102cd7e <pvclock_resume+0>:  movq   $0x0,0x5a42cf(%rip)        # 0xffffffff815d1058
0xffffffff8102cd89 <pvclock_resume+11>: retq

crash> x/x 0xffffffff815d1058
0xffffffff815d1058:     0x000001a4254e0867

crash> p/x last_value
$17 = {
  counter = 0x1a4254e0867
}



crash> ps | grep -v IN
   PID    PPID  CPU       TASK        ST  %MEM     VSZ    RSS  COMM
>     0      0   0  ffffffff814611f0  RU   0.0       0      0  [swapper]
>     0      0   1  ffff88007ff50e20  RU   0.0       0      0  [swapper]
>     0      0   2  ffff88007ff51530  RU   0.0       0      0  [swapper]
>     0      0   3  ffff88007ff51c40  RU   0.0       0      0  [swapper]
>     0      0   4  ffff88007ff52350  RU   0.0       0      0  [swapper]
>     0      0   5  ffff88007ff52a60  RU   0.0       0      0  [swapper]
     21      2   0  ffff88007ffab170  UN   0.0       0      0  [events/0]
     26      2   5  ffff88007ffad4c0  UN   0.0       0      0  [events/5]
     32      2   5  ffff88007f418000  UN   0.0       0      0  [xenwatch]
crash> bt 0 21 26 32
PID: 0      TASK: ffffffff814611f0  CPU: 0   COMMAND: "swapper"
 #0 [ffffffff8142df70] xen_safe_halt at ffffffff8100dcbf
 #1 [ffffffff8142df78] xen_idle at ffffffff8100be63
 #2 [ffffffff8142df90] cpu_idle at ffffffff8100fe97

PID: 0      TASK: ffff88007ff50e20  CPU: 1   COMMAND: "swapper"
 #0 [ffff88007ff5de50] schedule at ffffffff812fb2a7
 #1 [ffff88007ff5de68] xen_force_evtchn_callback at ffffffff8100dc41
 #2 [ffff88007ff5de70] check_events at ffffffff8100e252
 #3 [ffff88007ff5dec8] tick_nohz_stop_sched_tick at ffffffff81070d4e
 #4 [ffff88007ff5df28] cpu_idle at ffffffff8100fe97

PID: 0      TASK: ffff88007ff51530  CPU: 2   COMMAND: "swapper"
 #0 [ffff88007ff5fe50] schedule at ffffffff812fb2a7
 #1 [ffff88007ff5fe68] xen_force_evtchn_callback at ffffffff8100dc41
 #2 [ffff88007ff5fe70] check_events at ffffffff8100e252
 #3 [ffff88007ff5fec8] tick_nohz_stop_sched_tick at ffffffff81070d4e
 #4 [ffff88007ff5ff28] cpu_idle at ffffffff8100fe97

PID: 0      TASK: ffff88007ff51c40  CPU: 3   COMMAND: "swapper"
 #0 [ffff88007ff69e50] schedule at ffffffff812fb2a7
 #1 [ffff88007ff69e68] xen_force_evtchn_callback at ffffffff8100dc41
 #2 [ffff88007ff69e70] check_events at ffffffff8100e252
 #3 [ffff88007ff69ec8] tick_nohz_stop_sched_tick at ffffffff81070d4e
 #4 [ffff88007ff69f28] cpu_idle at ffffffff8100fe97

PID: 0      TASK: ffff88007ff52350  CPU: 4   COMMAND: "swapper"
 #0 [ffff88007ff6be50] schedule at ffffffff812fb2a7
 #1 [ffff88007ff6be68] xen_force_evtchn_callback at ffffffff8100dc41
 #2 [ffff88007ff6be70] check_events at ffffffff8100e252
 #3 [ffff88007ff6bec8] tick_nohz_stop_sched_tick at ffffffff81070d4e
 #4 [ffff88007ff6bf28] cpu_idle at ffffffff8100fe97

PID: 0      TASK: ffff88007ff52a60  CPU: 5   COMMAND: "swapper"
 #0 [ffff88007ff6de50] schedule at ffffffff812fb2a7
 #1 [ffff88007ff6de68] xen_force_evtchn_callback at ffffffff8100dc41
 #2 [ffff88007ff6de70] check_events at ffffffff8100e252
 #3 [ffff88007ff6dec8] tick_nohz_stop_sched_tick at ffffffff81070d4e
 #4 [ffff88007ff6df28] cpu_idle at ffffffff8100fe97

PID: 21     TASK: ffff88007ffab170  CPU: 0   COMMAND: "events/0"
 #0 [ffff88007ffc3b90] schedule at ffffffff812fb2a7
 #1 [ffff88007ffc3c68] schedule_timeout at ffffffff812fb6dd
 #2 [ffff88007ffc3ce8] wait_for_common at ffffffff812fb594
 #3 [ffff88007ffc3d78] synchronize_sched at ffffffff8106307b
 #4 [ffff88007ffc3db8] dev_deactivate at ffffffff81262d5f
 #5 [ffff88007ffc3de8] __linkwatch_run_queue at ffffffff8125a8ea
 #6 [ffff88007ffc3e28] linkwatch_event at ffffffff8125a954
 #7 [ffff88007ffc3e38] worker_thread at ffffffff8106195f
 #8 [ffff88007ffc3ee8] kthread at ffffffff81064cc5
 #9 [ffff88007ffc3f48] kernel_thread at ffffffff81011baa

PID: 26     TASK: ffff88007ffad4c0  CPU: 5   COMMAND: "events/5"
 #0 [ffff88007ffd5cc0] schedule at ffffffff812fb2a7
 #1 [ffff88007ffd5d98] __mutex_lock_common at ffffffff812fbb3b
 #2 [ffff88007ffd5e08] mutex_lock at ffffffff812fbc63
 #3 [ffff88007ffd5e28] linkwatch_event at ffffffff8125a93d
 #4 [ffff88007ffd5e38] worker_thread at ffffffff8106195f
 #5 [ffff88007ffd5ee8] kthread at ffffffff81064cc5
 #6 [ffff88007ffd5f48] kernel_thread at ffffffff81011baa

PID: 32     TASK: ffff88007f418000  CPU: 5   COMMAND: "xenwatch"
 #0 [ffff88007f413c60] schedule at ffffffff812fb2a7
 #1 [ffff88007f413d38] __mutex_lock_common at ffffffff812fbb3b
 #2 [ffff88007f413da8] mutex_lock at ffffffff812fbc63
 #3 [ffff88007f413dc8] netif_notify_peers at ffffffff8126316d
 #4 [ffff88007f413dd8] backend_changed at ffffffffa000a1b0
 #5 [ffff88007f413e78] xenwatch_thread at ffffffff811f1628
 #6 [ffff88007f413ee8] kthread at ffffffff81064cc5
 #7 [ffff88007f413f48] kernel_thread at ffffffff81011baa

-Timo



Reply to: