lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4790D187.8050000@goop.org>
Date:	Fri, 18 Jan 2008 08:19:19 -0800
From:	Jeremy Fitzhardinge <jeremy@...p.org>
To:	xming <xmingske@...il.com>
CC:	linux-kernel@...r.kernel.org,
	Xen-devel <xen-devel@...ts.xensource.com>
Subject: Re: Cannot boot xen DomU > 2.6.23.1

xming wrote:
>> OK, I misunderstood your original report to mean that something was
>> complaining about "too much" output.  You're saying that lots of console
>> output seems to lock the domain.
>>     
>
> Sorry about that, and yes that is the case.
>
>   
>> I've had a report about heavy disk IO seems to lock up as well.  Perhaps
>> they're both related to high event rates.  Do you think you could try an
>> IO-intensive workload to see if you can get a similar lockup?
>>     
>
> IO-intensive locks up too (see below)
>
>   
>> When the domain is locked up, what does /usr/lib/xen/bin/xenctx say?
>>     
>
> see below
>
>   
>> Hm.  Rather than backing out the structure-change patch, could you try
>> this workaround:
>>
>> diff -r be3ca4e0e19e arch/x86/xen/enlighten.c
>> --- a/arch/x86/xen/enlighten.c  Thu Jan 17 14:25:07 2008 -0800
>> +++ b/arch/x86/xen/enlighten.c  Thu Jan 17 16:37:42 2008 -0800
>> @@ -95,7 +95,7 @@ struct shared_info *HYPERVISOR_shared_in
>>   *
>>   * 0: not available, 1: available
>>   */
>> -static int have_vcpu_info_placement = 1;
>> +static int have_vcpu_info_placement = 0;
>>
>>  static void __init xen_vcpu_setup(int cpu)
>>  {
>>     
>
> First of all this patch solves the lock-ups, it works as advertised :)

OK, good.  I guess events are getting lost somewhere with vcpu_info 
placement.

> # /usr/lib/xen/bin/xenctx 113
>   

Would it be possible to map the eip and some top parts of the stack back 
to kernel symbols?  Seems to be the same place in both traces, which is 
interesting.

> eip: c037c0c7
> esp: c0343f90
> eax: 00000000   ebx: 00000001   ecx: 00000000   edx: c0342000
> esi: c0373004   edi: c1210df4   ebp: 00001b7d
>  cs: 00000061    ds: 0000007b    fs: 000000d8    gs: 00000000
>
> Stack:
>  c0100add c0378980 c0101962 c0104821 c120a000 c0378df4 c0348cff 00000025
>  c0348430 00000004 00009000 00006df4 00ea1000 c0363be0 c0343fe8 c03dd007
>  00000000 c0343fec c0349868 c0343fe0 178bc1f1 00002001 00020800 00060fb1
>  00000000 c03dd000 00000000 00000000
>
> Code:
> cc cc cc cc cc cc cc cc cc cc cc cc cc cc b8 06 00 00 00 cd 82 <c3> cc
> cc cc cc cc cc cc cc cc cc
>
> Call Trace:
>   [<c037c0c7>]  <--
>   [<c0100add>]
>   [<c0378980>]
>   [<c0101962>]
>   [<c0104821>]
>   [<c120a000>]
>   [<c0378df4>]
>   [<c0348cff>]
>   [<c0348430>]
>   [<c0363be0>]
>   [<c0343fe8>]
>   [<c03dd007>]
>   [<c0343fec>]
>   [<c0349868>]
>   [<c0343fe0>]
>   [<178bc1f1>]
>   [<c03dd000>]
>
> Scenario 2 (have_vcpu_info_placement = 0)
> --------------------------------------------------------------
>
> test1: no crash
> test2: no crash, but occationally I still get funny output like this
>
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 0000AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 000AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 000AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
> 00AAAAAAAAAAAAAAAAAAAAAAAAAAZZ
>   

Hm, I guess some of the output is getting dropped.  Does this happen 
with 2.6.18-xen?

    J
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ