debuggers.hg
changeset 3770:d21fbb46b9d8
bitkeeper revision 1.1159.253.1 (4208f8a54Zaz-XgC11YTHeLxPHPoZg)
Merge scramble.cl.cam.ac.uk:/auto/groups/xeno/BK/xeno.bk
into scramble.cl.cam.ac.uk:/local/scratch/kaf24/xen-unstable.bk
Merge scramble.cl.cam.ac.uk:/auto/groups/xeno/BK/xeno.bk
into scramble.cl.cam.ac.uk:/local/scratch/kaf24/xen-unstable.bk
author | kaf24@scramble.cl.cam.ac.uk |
---|---|
date | Tue Feb 08 17:36:37 2005 +0000 (2005-02-08) |
parents | f5f2757b3aa2 cb87fd290eb0 |
children | 12104922e743 |
files | xen/arch/x86/extable.c xen/arch/x86/mm.c xen/arch/x86/shadow.c xen/arch/x86/traps.c xen/arch/x86/vmx.c xen/common/dom_mem_ops.c xen/common/page_alloc.c xen/common/physdev.c xen/include/asm-x86/mm.h xen/include/asm-x86/shadow.h xen/include/asm-x86/x86_64/regs.h |
line diff
1.1 --- a/xen/arch/x86/extable.c Tue Feb 08 16:44:16 2005 +0000 1.2 +++ b/xen/arch/x86/extable.c Tue Feb 08 17:36:37 2005 +0000 1.3 @@ -67,6 +67,6 @@ search_pre_exception_table(unsigned long 1.4 { 1.5 unsigned long fixup = search_one_table( 1.6 __start___pre_ex_table, __stop___pre_ex_table-1, addr); 1.7 - DPRINTK("Pre-exception: %08lx -> %08lx\n", addr, fixup); 1.8 + DPRINTK("Pre-exception: %p -> %p\n", addr, fixup); 1.9 return fixup; 1.10 }
2.1 --- a/xen/arch/x86/mm.c Tue Feb 08 16:44:16 2005 +0000 2.2 +++ b/xen/arch/x86/mm.c Tue Feb 08 17:36:37 2005 +0000 2.3 @@ -298,13 +298,13 @@ static int get_page_from_pagenr(unsigned 2.4 2.5 if ( unlikely(!pfn_is_ram(page_nr)) ) 2.6 { 2.7 - MEM_LOG("Pfn %08lx is not RAM", page_nr); 2.8 + MEM_LOG("Pfn %p is not RAM", page_nr); 2.9 return 0; 2.10 } 2.11 2.12 if ( unlikely(!get_page(page, d)) ) 2.13 { 2.14 - MEM_LOG("Could not get page ref for pfn %08lx", page_nr); 2.15 + MEM_LOG("Could not get page ref for pfn %p", page_nr); 2.16 return 0; 2.17 } 2.18 2.19 @@ -323,11 +323,9 @@ static int get_page_and_type_from_pagenr 2.20 2.21 if ( unlikely(!get_page_type(page, type)) ) 2.22 { 2.23 -#ifdef VERBOSE 2.24 if ( (type & PGT_type_mask) != PGT_l1_page_table ) 2.25 - MEM_LOG("Bad page type for pfn %08lx (%08x)", 2.26 + MEM_LOG("Bad page type for pfn %p (%08x)", 2.27 page_nr, page->u.inuse.type_info); 2.28 -#endif 2.29 put_page(page); 2.30 return 0; 2.31 } 2.32 @@ -420,7 +418,7 @@ get_page_from_l1e( 2.33 if ( IS_CAPABLE_PHYSDEV(d) ) 2.34 return domain_iomem_in_pfn(d, pfn); 2.35 2.36 - MEM_LOG("Non-privileged attempt to map I/O space %08lx", pfn); 2.37 + MEM_LOG("Non-privileged attempt to map I/O space %p", pfn); 2.38 return 0; 2.39 } 2.40 2.41 @@ -617,7 +615,7 @@ static inline int update_l2e(l2_pgentry_ 2.42 l2_pgentry_val(ol2e), 2.43 l2_pgentry_val(nl2e)); 2.44 if ( o != l2_pgentry_val(ol2e) ) 2.45 - MEM_LOG("Failed to update %08lx -> %08lx: saw %08lx\n", 2.46 + MEM_LOG("Failed to update %p -> %p: saw %p\n", 2.47 l2_pgentry_val(ol2e), l2_pgentry_val(nl2e), o); 2.48 return (o == l2_pgentry_val(ol2e)); 2.49 } 2.50 @@ -681,7 +679,7 @@ static inline int update_l1e(l1_pgentry_ 2.51 if ( unlikely(cmpxchg_user(pl1e, o, n) != 0) || 2.52 unlikely(o != l1_pgentry_val(ol1e)) ) 2.53 { 2.54 - MEM_LOG("Failed to update %08lx -> %08lx: saw %08lx\n", 2.55 + MEM_LOG("Failed to update %p -> %p: saw %p\n", 2.56 l1_pgentry_val(ol1e), l1_pgentry_val(nl1e), o); 2.57 return 0; 2.58 } 2.59 @@ -845,7 +843,7 @@ int get_page_type(struct pfn_info *page, 2.60 nx = x + 1; 2.61 if ( unlikely((nx & PGT_count_mask) == 0) ) 2.62 { 2.63 - MEM_LOG("Type count overflow on pfn %08lx\n", page_to_pfn(page)); 2.64 + MEM_LOG("Type count overflow on pfn %p\n", page_to_pfn(page)); 2.65 return 0; 2.66 } 2.67 else if ( unlikely((x & PGT_count_mask) == 0) ) 2.68 @@ -881,7 +879,7 @@ int get_page_type(struct pfn_info *page, 2.69 { 2.70 if ( ((x & PGT_type_mask) != PGT_l2_page_table) || 2.71 ((type & PGT_type_mask) != PGT_l1_page_table) ) 2.72 - MEM_LOG("Bad type (saw %08x != exp %08x) for pfn %08lx\n", 2.73 + MEM_LOG("Bad type (saw %08x != exp %08x) for pfn %p\n", 2.74 x & PGT_type_mask, type, page_to_pfn(page)); 2.75 return 0; 2.76 } 2.77 @@ -916,7 +914,7 @@ int get_page_type(struct pfn_info *page, 2.78 /* Try to validate page type; drop the new reference on failure. */ 2.79 if ( unlikely(!alloc_page_type(page, type & PGT_type_mask)) ) 2.80 { 2.81 - MEM_LOG("Error while validating pfn %08lx for type %08x." 2.82 + MEM_LOG("Error while validating pfn %p for type %08x." 2.83 " caf=%08x taf=%08x\n", 2.84 page_to_pfn(page), type, 2.85 page->count_info, 2.86 @@ -958,7 +956,7 @@ int new_guest_cr3(unsigned long pfn) 2.87 } 2.88 else 2.89 { 2.90 - MEM_LOG("Error while installing new baseptr %08lx", pfn); 2.91 + MEM_LOG("Error while installing new baseptr %p", pfn); 2.92 } 2.93 2.94 return okay; 2.95 @@ -993,14 +991,14 @@ static int do_extended_command(unsigned 2.96 2.97 if ( unlikely(!okay) ) 2.98 { 2.99 - MEM_LOG("Error while pinning pfn %08lx", pfn); 2.100 + MEM_LOG("Error while pinning pfn %p", pfn); 2.101 break; 2.102 } 2.103 2.104 if ( unlikely(test_and_set_bit(_PGT_pinned, 2.105 &page->u.inuse.type_info)) ) 2.106 { 2.107 - MEM_LOG("Pfn %08lx already pinned", pfn); 2.108 + MEM_LOG("Pfn %p already pinned", pfn); 2.109 put_page_and_type(page); 2.110 okay = 0; 2.111 break; 2.112 @@ -1011,7 +1009,7 @@ static int do_extended_command(unsigned 2.113 case MMUEXT_UNPIN_TABLE: 2.114 if ( unlikely(!(okay = get_page_from_pagenr(pfn, FOREIGNDOM))) ) 2.115 { 2.116 - MEM_LOG("Page %08lx bad domain (dom=%p)", 2.117 + MEM_LOG("Page %p bad domain (dom=%p)", 2.118 ptr, page_get_owner(page)); 2.119 } 2.120 else if ( likely(test_and_clear_bit(_PGT_pinned, 2.121 @@ -1024,7 +1022,7 @@ static int do_extended_command(unsigned 2.122 { 2.123 okay = 0; 2.124 put_page(page); 2.125 - MEM_LOG("Pfn %08lx not pinned", pfn); 2.126 + MEM_LOG("Pfn %p not pinned", pfn); 2.127 } 2.128 break; 2.129 2.130 @@ -1061,7 +1059,7 @@ static int do_extended_command(unsigned 2.131 ((ptr+ents*LDT_ENTRY_SIZE) > PAGE_OFFSET) ) 2.132 { 2.133 okay = 0; 2.134 - MEM_LOG("Bad args to SET_LDT: ptr=%08lx, ents=%08lx", ptr, ents); 2.135 + MEM_LOG("Bad args to SET_LDT: ptr=%p, ents=%p", ptr, ents); 2.136 } 2.137 else if ( (ed->arch.ldt_ents != ents) || 2.138 (ed->arch.ldt_base != ptr) ) 2.139 @@ -1130,7 +1128,7 @@ static int do_extended_command(unsigned 2.140 unlikely(!pfn_is_ram(pfn)) || 2.141 unlikely((e = find_domain_by_id(domid)) == NULL) ) 2.142 { 2.143 - MEM_LOG("Bad frame (%08lx) or bad domid (%d).\n", pfn, domid); 2.144 + MEM_LOG("Bad frame (%p) or bad domid (%d).\n", pfn, domid); 2.145 okay = 0; 2.146 break; 2.147 } 2.148 @@ -1150,7 +1148,7 @@ static int do_extended_command(unsigned 2.149 (1|PGC_allocated)) || 2.150 unlikely(nd != d) ) 2.151 { 2.152 - MEM_LOG("Bad page values %08lx: ed=%p(%u), sd=%p," 2.153 + MEM_LOG("Bad page values %p: ed=%p(%u), sd=%p," 2.154 " caf=%08x, taf=%08x\n", page_to_pfn(page), 2.155 d, d->id, nd, x, page->u.inuse.type_info); 2.156 spin_unlock(&d->page_alloc_lock); 2.157 @@ -1186,7 +1184,7 @@ static int do_extended_command(unsigned 2.158 unlikely(!gnttab_prepare_for_transfer(e, d, gntref)) ) 2.159 { 2.160 MEM_LOG("Transferee has no reservation headroom (%d,%d), or " 2.161 - "provided a bad grant ref, or is dying (%08lx).\n", 2.162 + "provided a bad grant ref, or is dying (%p).\n", 2.163 e->tot_pages, e->max_pages, e->d_flags); 2.164 spin_unlock(&e->page_alloc_lock); 2.165 put_domain(e); 2.166 @@ -1219,7 +1217,7 @@ static int do_extended_command(unsigned 2.167 e = percpu_info[cpu].foreign; 2.168 if ( unlikely(e == NULL) ) 2.169 { 2.170 - MEM_LOG("No FOREIGNDOM to reassign pfn %08lx to", pfn); 2.171 + MEM_LOG("No FOREIGNDOM to reassign pfn %p to", pfn); 2.172 okay = 0; 2.173 break; 2.174 } 2.175 @@ -1262,7 +1260,7 @@ static int do_extended_command(unsigned 2.176 (1|PGC_allocated)) || 2.177 unlikely(nd != d) ) 2.178 { 2.179 - MEM_LOG("Bad page values %08lx: ed=%p(%u), sd=%p," 2.180 + MEM_LOG("Bad page values %p: ed=%p(%u), sd=%p," 2.181 " caf=%08x, taf=%08x\n", page_to_pfn(page), 2.182 d, d->id, nd, x, page->u.inuse.type_info); 2.183 okay = 0; 2.184 @@ -1304,7 +1302,7 @@ static int do_extended_command(unsigned 2.185 break; 2.186 2.187 default: 2.188 - MEM_LOG("Invalid extended pt command 0x%08lx", val & MMUEXT_CMD_MASK); 2.189 + MEM_LOG("Invalid extended pt command 0x%p", val & MMUEXT_CMD_MASK); 2.190 okay = 0; 2.191 break; 2.192 } 2.193 @@ -1511,7 +1509,7 @@ int do_mmu_update( 2.194 break; 2.195 2.196 default: 2.197 - MEM_LOG("Invalid page update command %08lx", req.ptr); 2.198 + MEM_LOG("Invalid page update command %p", req.ptr); 2.199 break; 2.200 } 2.201 2.202 @@ -1876,7 +1874,7 @@ void ptwr_flush(const int which) 2.203 */ 2.204 BUG(); 2.205 } 2.206 - PTWR_PRINTK("[%c] disconnected_l1va at %p is %08lx\n", 2.207 + PTWR_PRINTK("[%c] disconnected_l1va at %p is %p\n", 2.208 PTWR_PRINT_WHICH, ptep, pte); 2.209 pte &= ~_PAGE_RW; 2.210 2.211 @@ -1911,7 +1909,7 @@ void ptwr_flush(const int which) 2.212 #else 2.213 flush_tlb_all(); 2.214 #endif 2.215 - PTWR_PRINTK("[%c] disconnected_l1va at %p now %08lx\n", 2.216 + PTWR_PRINTK("[%c] disconnected_l1va at %p now %p\n", 2.217 PTWR_PRINT_WHICH, ptep, pte); 2.218 2.219 /* 2.220 @@ -2035,7 +2033,7 @@ int ptwr_do_page_fault(unsigned long add 2.221 2.222 if ( l2_idx == (addr >> L2_PAGETABLE_SHIFT) ) 2.223 { 2.224 - MEM_LOG("PTWR failure! Pagetable maps itself at %08lx\n", addr); 2.225 + MEM_LOG("PTWR failure! Pagetable maps itself at %p\n", addr); 2.226 domain_crash(); 2.227 } 2.228 2.229 @@ -2064,8 +2062,8 @@ int ptwr_do_page_fault(unsigned long add 2.230 } 2.231 } 2.232 2.233 - PTWR_PRINTK("[%c] page_fault on l1 pt at va %08lx, pt for %08x, " 2.234 - "pfn %08lx\n", PTWR_PRINT_WHICH, 2.235 + PTWR_PRINTK("[%c] page_fault on l1 pt at va %p, pt for %08x, " 2.236 + "pfn %p\n", PTWR_PRINT_WHICH, 2.237 addr, l2_idx << L2_PAGETABLE_SHIFT, pfn); 2.238 2.239 /* 2.240 @@ -2098,7 +2096,7 @@ int ptwr_do_page_fault(unsigned long add 2.241 2.242 /* Finally, make the p.t. page writable by the guest OS. */ 2.243 pte |= _PAGE_RW; 2.244 - PTWR_PRINTK("[%c] update %p pte to %08lx\n", PTWR_PRINT_WHICH, 2.245 + PTWR_PRINTK("[%c] update %p pte to %p\n", PTWR_PRINT_WHICH, 2.246 &linear_pg_table[addr>>PAGE_SHIFT], pte); 2.247 if ( unlikely(__put_user(pte, (unsigned long *) 2.248 &linear_pg_table[addr>>PAGE_SHIFT])) ) 2.249 @@ -2157,7 +2155,7 @@ void ptwr_status(void) 2.250 page = &frame_table[pfn]; 2.251 printk("need to alloc l1 page %p\n", page); 2.252 /* make pt page writable */ 2.253 - printk("need to make read-only l1-page at %p is %08lx\n", 2.254 + printk("need to make read-only l1-page at %p is %p\n", 2.255 ptep, pte); 2.256 2.257 if ( ptwr_info[cpu].ptinfo[PTWR_PT_ACTIVE].l1va == 0 )
3.1 --- a/xen/arch/x86/shadow.c Tue Feb 08 16:44:16 2005 +0000 3.2 +++ b/xen/arch/x86/shadow.c Tue Feb 08 17:36:37 2005 +0000 3.3 @@ -249,7 +249,7 @@ static int shadow_mode_table_op( 3.4 3.5 ASSERT(spin_is_locked(&d->arch.shadow_lock)); 3.6 3.7 - SH_VLOG("shadow mode table op %08lx %08lx count %d", 3.8 + SH_VLOG("shadow mode table op %p %p count %d", 3.9 pagetable_val(d->exec_domain[0]->arch.pagetable), /* XXX SMP */ 3.10 pagetable_val(d->exec_domain[0]->arch.shadow_table), /* XXX SMP */ 3.11 d->arch.shadow_page_count); 3.12 @@ -422,7 +422,7 @@ void unshadow_table(unsigned long gpfn, 3.13 unsigned long smfn; 3.14 struct domain *d = page_get_owner(&frame_table[gpfn]); 3.15 3.16 - SH_VLOG("unshadow_table type=%08x gpfn=%08lx", type, gpfn); 3.17 + SH_VLOG("unshadow_table type=%08x gpfn=%p", type, gpfn); 3.18 3.19 perfc_incrc(unshadow_table_count); 3.20 3.21 @@ -455,7 +455,7 @@ unsigned long shadow_l2_table( 3.22 3.23 guest_gpfn = __mfn_to_gpfn(d, gpfn); 3.24 3.25 - SH_VVLOG("shadow_l2_table( %08lx )", gpfn); 3.26 + SH_VVLOG("shadow_l2_table( %p )", gpfn); 3.27 3.28 perfc_incrc(shadow_l2_table_count); 3.29 3.30 @@ -506,7 +506,7 @@ unsigned long shadow_l2_table( 3.31 if ( shadow_mode(d) != SHM_full_32 ) 3.32 unmap_domain_mem(spl2e); 3.33 3.34 - SH_VLOG("shadow_l2_table( %08lx -> %08lx)", gpfn, spfn); 3.35 + SH_VLOG("shadow_l2_table( %p -> %p)", gpfn, spfn); 3.36 return spfn; 3.37 } 3.38 3.39 @@ -526,7 +526,7 @@ static void shadow_map_l1_into_current_l 3.40 if ( !(sl1ss & PSH_shadowed) ) 3.41 { 3.42 /* This L1 is NOT already shadowed so we need to shadow it. */ 3.43 - SH_VVLOG("4a: l1 not shadowed ( %08lx )", sl1pfn); 3.44 + SH_VVLOG("4a: l1 not shadowed ( %p )", sl1pfn); 3.45 3.46 sl1pfn_info = alloc_shadow_page(d); 3.47 sl1pfn_info->u.inuse.type_info = PGT_l1_page_table; 3.48 @@ -555,7 +555,7 @@ static void shadow_map_l1_into_current_l 3.49 else 3.50 { 3.51 /* This L1 is shadowed already, but the L2 entry is missing. */ 3.52 - SH_VVLOG("4b: was shadowed, l2 missing ( %08lx )", sl1pfn); 3.53 + SH_VVLOG("4b: was shadowed, l2 missing ( %p )", sl1pfn); 3.54 3.55 sl1pfn = sl1ss & PSH_pfn_mask; 3.56 l2pde_general(d, &gl2e, &sl2e, sl1pfn); 3.57 @@ -596,7 +596,7 @@ int shadow_fault(unsigned long va, long 3.58 struct exec_domain *ed = current; 3.59 struct domain *d = ed->domain; 3.60 3.61 - SH_VVLOG("shadow_fault( va=%08lx, code=%ld )", va, error_code ); 3.62 + SH_VVLOG("shadow_fault( va=%p, code=%ld )", va, error_code ); 3.63 3.64 check_pagetable(d, ed->arch.pagetable, "pre-sf"); 3.65 3.66 @@ -678,7 +678,7 @@ int shadow_fault(unsigned long va, long 3.67 if ( unlikely(__put_user(spte, (unsigned long *) 3.68 &shadow_linear_pg_table[va >> PAGE_SHIFT])) ) 3.69 { 3.70 - SH_VVLOG("3: not shadowed/mapped gpte=%08lx spte=%08lx", gpte, spte); 3.71 + SH_VVLOG("3: not shadowed/mapped gpte=%p spte=%p", gpte, spte); 3.72 shadow_map_l1_into_current_l2(va); 3.73 shadow_linear_pg_table[va >> PAGE_SHIFT] = mk_l1_pgentry(spte); 3.74 } 3.75 @@ -702,8 +702,8 @@ void shadow_l1_normal_pt_update( 3.76 l1_pgentry_t *spl1e, *prev_spl1e = *prev_spl1e_ptr; 3.77 3.78 /* N.B. To get here, we know the l1 page *must* be shadowed. */ 3.79 - SH_VVLOG("shadow_l1_normal_pt_update pa=%08lx, gpte=%08lx, " 3.80 - "prev_smfn=%08lx, prev_spl1e=%p", 3.81 + SH_VVLOG("shadow_l1_normal_pt_update pa=%p, gpte=%p, " 3.82 + "prev_smfn=%p, prev_spl1e=%p", 3.83 pa, gpte, prev_smfn, prev_spl1e); 3.84 3.85 smfn = __shadow_status(current->domain, pa >> PAGE_SHIFT) & PSH_pfn_mask; 3.86 @@ -732,7 +732,7 @@ void shadow_l2_normal_pt_update(unsigned 3.87 unsigned long sl1mfn; 3.88 3.89 /* N.B. To get here, we know the l2 page *must* be shadowed. */ 3.90 - SH_VVLOG("shadow_l2_normal_pt_update pa=%08lx, gpde=%08lx",pa,gpde); 3.91 + SH_VVLOG("shadow_l2_normal_pt_update pa=%p, gpde=%p",pa,gpde); 3.92 3.93 sl2mfn = __shadow_status(current->domain, pa >> PAGE_SHIFT) & PSH_pfn_mask; 3.94 3.95 @@ -771,8 +771,8 @@ char * sh_check_name; 3.96 3.97 #define FAIL(_f, _a...) \ 3.98 do { \ 3.99 - printk("XXX %s-FAIL (%d,%d)" _f " g=%08lx s=%08lx &g=%08lx &s=%08lx" \ 3.100 - " pa(&g)=%08lx pa(&s)=%08lx\n", \ 3.101 + printk("XXX %s-FAIL (%d,%d)" _f " g=%p s=%p &g=%p &s=%p" \ 3.102 + " pa(&g)=%p pa(&s)=%p\n", \ 3.103 sh_check_name, level, i, ## _a , gpte, spte, pgpte, pspte, \ 3.104 virt_to_phys2(pgpte), virt_to_phys2(pspte)); \ 3.105 errors++; \ 3.106 @@ -829,7 +829,7 @@ static int check_pte( 3.107 FAIL("Shadow in L1 entry?"); 3.108 3.109 if ( __shadow_status(d, gpfn) != (PSH_shadowed | smfn) ) 3.110 - FAIL("smfn problem g.sf=%08lx", 3.111 + FAIL("smfn problem g.sf=%p", 3.112 __shadow_status(d, gpfn) ); 3.113 } 3.114 3.115 @@ -885,7 +885,7 @@ void check_pagetable(struct domain *d, p 3.116 3.117 if ( ! (ss & PSH_shadowed) ) 3.118 { 3.119 - printk("%s-PT %08lx not shadowed\n", s, gptbase); 3.120 + printk("%s-PT %p not shadowed\n", s, gptbase); 3.121 3.122 if ( ss != 0 ) 3.123 BUG(); 3.124 @@ -908,7 +908,7 @@ void check_pagetable(struct domain *d, p 3.125 for ( i = DOMAIN_ENTRIES_PER_L2_PAGETABLE; 3.126 i < (SH_LINEAR_PT_VIRT_START >> L2_PAGETABLE_SHIFT); 3.127 i++ ) 3.128 - printk("+++ (%d) %08lx %08lx\n",i, 3.129 + printk("+++ (%d) %p %p\n",i, 3.130 l2_pgentry_val(gpl2e[i]), l2_pgentry_val(spl2e[i])); 3.131 FAILPT("hypervisor entries inconsistent"); 3.132 } 3.133 @@ -920,7 +920,7 @@ void check_pagetable(struct domain *d, p 3.134 if ( (l2_pgentry_val(spl2e[SH_LINEAR_PT_VIRT_START >> 3.135 L2_PAGETABLE_SHIFT]) != 3.136 ((smfn << PAGE_SHIFT) | __PAGE_HYPERVISOR)) ) 3.137 - FAILPT("hypervisor shadow linear map inconsistent %08lx %08lx", 3.138 + FAILPT("hypervisor shadow linear map inconsistent %p %p", 3.139 l2_pgentry_val(spl2e[SH_LINEAR_PT_VIRT_START >> 3.140 L2_PAGETABLE_SHIFT]), 3.141 (smfn << PAGE_SHIFT) | __PAGE_HYPERVISOR);
4.1 --- a/xen/arch/x86/traps.c Tue Feb 08 16:44:16 2005 +0000 4.2 +++ b/xen/arch/x86/traps.c Tue Feb 08 17:36:37 2005 +0000 4.3 @@ -272,12 +272,6 @@ asmlinkage int do_page_fault(struct xen_ 4.4 4.5 perfc_incrc(page_faults); 4.6 4.7 -#if 0 4.8 - printk("do_page_fault(addr=0x%08lx, error_code=%d)\n", 4.9 - addr, regs->error_code); 4.10 - show_registers(regs); 4.11 -#endif 4.12 - 4.13 if ( likely(VM_ASSIST(d, VMASST_TYPE_writable_pagetables)) ) 4.14 { 4.15 LOCK_BIGLOCK(d);
5.1 --- a/xen/arch/x86/vmx.c Tue Feb 08 16:44:16 2005 +0000 5.2 +++ b/xen/arch/x86/vmx.c Tue Feb 08 17:36:37 2005 +0000 5.3 @@ -277,7 +277,7 @@ static void vmx_vmexit_do_invlpg(unsigne 5.4 5.5 __vmread(GUEST_EIP, &eip); 5.6 5.7 - VMX_DBG_LOG(DBG_LEVEL_VMMU, "vmx_vmexit_do_invlpg:eip=%08lx, va=%08lx\n", 5.8 + VMX_DBG_LOG(DBG_LEVEL_VMMU, "vmx_vmexit_do_invlpg:eip=%p, va=%p\n", 5.9 eip, va); 5.10 5.11 /* 5.12 @@ -339,7 +339,7 @@ static void vmx_io_instruction(struct xe 5.13 __vmread(GUEST_EIP, &eip); 5.14 5.15 VMX_DBG_LOG(DBG_LEVEL_1, 5.16 - "vmx_io_instruction: eip=%08lx, exit_qualification = %lx\n", 5.17 + "vmx_io_instruction: eip=%p, exit_qualification = %lx\n", 5.18 eip, exit_qualification); 5.19 5.20 if (test_bit(6, &exit_qualification)) 5.21 @@ -653,7 +653,7 @@ static inline void vmx_vmexit_do_hlt() 5.22 unsigned long eip; 5.23 __vmread(GUEST_EIP, &eip); 5.24 #endif 5.25 - VMX_DBG_LOG(DBG_LEVEL_1, "vmx_vmexit_do_hlt:eip=%08lx\n", eip); 5.26 + VMX_DBG_LOG(DBG_LEVEL_1, "vmx_vmexit_do_hlt:eip=%p\n", eip); 5.27 __enter_scheduler(); 5.28 } 5.29 5.30 @@ -663,7 +663,7 @@ static inline void vmx_vmexit_do_mwait() 5.31 unsigned long eip; 5.32 __vmread(GUEST_EIP, &eip); 5.33 #endif 5.34 - VMX_DBG_LOG(DBG_LEVEL_1, "vmx_vmexit_do_mwait:eip=%08lx\n", eip); 5.35 + VMX_DBG_LOG(DBG_LEVEL_1, "vmx_vmexit_do_mwait:eip=%p\n", eip); 5.36 __enter_scheduler(); 5.37 } 5.38 5.39 @@ -915,7 +915,7 @@ asmlinkage void vmx_vmexit_handler(struc 5.40 break; 5.41 case EXIT_REASON_MSR_WRITE: 5.42 __vmread(GUEST_EIP, &eip); 5.43 - VMX_DBG_LOG(DBG_LEVEL_1, "MSR_WRITE: eip=%08lx, eax=%08lx, edx=%08lx", 5.44 + VMX_DBG_LOG(DBG_LEVEL_1, "MSR_WRITE: eip=%p, eax=%p, edx=%p", 5.45 eip, regs.eax, regs.edx); 5.46 /* just ignore this point */ 5.47 __get_instruction_length(inst_len);
6.1 --- a/xen/common/dom_mem_ops.c Tue Feb 08 16:44:16 2005 +0000 6.2 +++ b/xen/common/dom_mem_ops.c Tue Feb 08 17:36:37 2005 +0000 6.3 @@ -94,7 +94,7 @@ free_dom_mem(struct domain *d, 6.4 { 6.5 if ( unlikely((mpfn + j) >= max_page) ) 6.6 { 6.7 - DPRINTK("Domain %u page number out of range (%08lx>=%08lx)\n", 6.8 + DPRINTK("Domain %u page number out of range (%p>=%p)\n", 6.9 d->id, mpfn + j, max_page); 6.10 return i; 6.11 }
7.1 --- a/xen/common/page_alloc.c Tue Feb 08 16:44:16 2005 +0000 7.2 +++ b/xen/common/page_alloc.c Tue Feb 08 17:36:37 2005 +0000 7.3 @@ -165,7 +165,7 @@ void init_boot_pages(unsigned long ps, u 7.4 7.5 if ( (bad_pfn < (bitmap_size*8)) && !allocated_in_map(bad_pfn) ) 7.6 { 7.7 - printk("Marking page %08lx as bad\n", bad_pfn); 7.8 + printk("Marking page %p as bad\n", bad_pfn); 7.9 map_alloc(bad_pfn, 1); 7.10 } 7.11 }
8.1 --- a/xen/common/physdev.c Tue Feb 08 16:44:16 2005 +0000 8.2 +++ b/xen/common/physdev.c Tue Feb 08 17:36:37 2005 +0000 8.3 @@ -231,7 +231,7 @@ int domain_iomem_in_pfn(struct domain *p 8.4 phys_dev_t *phys_dev; 8.5 8.6 VERBOSE_INFO("Checking if physdev-capable domain %u needs access to " 8.7 - "pfn %08lx\n", p->id, pfn); 8.8 + "pfn %p\n", p->id, pfn); 8.9 8.10 spin_lock(&p->pcidev_lock); 8.11 8.12 @@ -257,7 +257,7 @@ int domain_iomem_in_pfn(struct domain *p 8.13 8.14 spin_unlock(&p->pcidev_lock); 8.15 8.16 - VERBOSE_INFO("Domain %u %s mapping of pfn %08lx\n", 8.17 + VERBOSE_INFO("Domain %u %s mapping of pfn %p\n", 8.18 p->id, ret ? "allowed" : "disallowed", pfn); 8.19 8.20 return ret;
9.1 --- a/xen/include/asm-x86/mm.h Tue Feb 08 16:44:16 2005 +0000 9.2 +++ b/xen/include/asm-x86/mm.h Tue Feb 08 17:36:37 2005 +0000 9.3 @@ -100,16 +100,13 @@ struct pfn_info 9.4 #define IS_XEN_HEAP_FRAME(_pfn) (page_to_phys(_pfn) < xenheap_phys_end) 9.5 9.6 #if defined(__i386__) 9.7 - 9.8 #define pickle_domptr(_d) ((u32)(unsigned long)(_d)) 9.9 #define unpickle_domptr(_d) ((struct domain *)(unsigned long)(_d)) 9.10 - 9.11 #elif defined(__x86_64__) 9.12 static inline struct domain *unpickle_domptr(u32 _domain) 9.13 { return (_domain == 0) ? NULL : __va(_domain); } 9.14 static inline u32 pickle_domptr(struct domain *domain) 9.15 { return (domain == NULL) ? 0 : (u32)__pa(domain); } 9.16 - 9.17 #endif 9.18 9.19 #define page_get_owner(_p) (unpickle_domptr((_p)->u.inuse._domain)) 9.20 @@ -171,7 +168,7 @@ static inline int get_page(struct pfn_in 9.21 unlikely((nx & PGC_count_mask) == 0) || /* Count overflow? */ 9.22 unlikely(d != _domain) ) /* Wrong owner? */ 9.23 { 9.24 - DPRINTK("Error pfn %08lx: ed=%p, sd=%p, caf=%08x, taf=%08x\n", 9.25 + DPRINTK("Error pfn %p: ed=%p, sd=%p, caf=%08x, taf=%08x\n", 9.26 page_to_pfn(page), domain, unpickle_domptr(d), 9.27 x, page->u.inuse.type_info); 9.28 return 0;
10.1 --- a/xen/include/asm-x86/shadow.h Tue Feb 08 16:44:16 2005 +0000 10.2 +++ b/xen/include/asm-x86/shadow.h Tue Feb 08 17:36:37 2005 +0000 10.3 @@ -237,7 +237,7 @@ static inline void l1pte_write_fault( 10.4 10.5 spte = (mfn << PAGE_SHIFT) | (gpte & ~PAGE_MASK); 10.6 10.7 - SH_VVLOG("l1pte_write_fault: updating spte=0x%08lx gpte=0x%08lx", spte, gpte); 10.8 + SH_VVLOG("l1pte_write_fault: updating spte=0x%p gpte=0x%p", spte, gpte); 10.9 *gpte_p = gpte; 10.10 *spte_p = spte; 10.11 } 10.12 @@ -256,7 +256,7 @@ static inline void l1pte_read_fault( 10.13 if ( (shadow_mode(d) == SHM_logdirty) || ! (gpte & _PAGE_DIRTY) ) 10.14 spte &= ~_PAGE_RW; 10.15 10.16 - SH_VVLOG("l1pte_read_fault: updating spte=0x%08lx gpte=0x%08lx", spte, gpte); 10.17 + SH_VVLOG("l1pte_read_fault: updating spte=0x%p gpte=0x%p", spte, gpte); 10.18 *gpte_p = gpte; 10.19 *spte_p = spte; 10.20 } 10.21 @@ -309,7 +309,7 @@ static inline void l1pte_propagate_from_ 10.22 10.23 #if SHADOW_VERBOSE_DEBUG 10.24 if ( old_spte || spte || gpte ) 10.25 - SH_VVLOG("l1pte_propagate_from_guest: gpte=0x%08lx, old spte=0x%08lx, new spte=0x%08lx ", gpte, old_spte, spte); 10.26 + SH_VVLOG("l1pte_propagate_from_guest: gpte=0x%p, old spte=0x%p, new spte=0x%p ", gpte, old_spte, spte); 10.27 #endif 10.28 10.29 *gpte_p = gpte; 10.30 @@ -366,7 +366,7 @@ static void shadow_audit(struct domain * 10.31 live++; 10.32 if ( (a->pfn == 0) || (a->smfn_and_flags == 0) ) 10.33 { 10.34 - printk("XXX live=%d pfn=%08lx sp=%08lx next=%p\n", 10.35 + printk("XXX live=%d pfn=%p sp=%p next=%p\n", 10.36 live, a->pfn, a->smfn_and_flags, a->next); 10.37 BUG(); 10.38 } 10.39 @@ -439,7 +439,7 @@ static inline unsigned long __shadow_sta 10.40 SWAP(head->smfn_and_flags, x->smfn_and_flags); 10.41 } 10.42 10.43 - SH_VVLOG("lookup gpfn=%08lx => status=%08lx", 10.44 + SH_VVLOG("lookup gpfn=%p => status=%p", 10.45 gpfn, head->smfn_and_flags); 10.46 return head->smfn_and_flags; 10.47 } 10.48 @@ -449,7 +449,7 @@ static inline unsigned long __shadow_sta 10.49 } 10.50 while ( x != NULL ); 10.51 10.52 - SH_VVLOG("lookup gpfn=%08lx => status=0", gpfn); 10.53 + SH_VVLOG("lookup gpfn=%p => status=0", gpfn); 10.54 return 0; 10.55 } 10.56 10.57 @@ -577,7 +577,7 @@ static inline void set_shadow_status( 10.58 10.59 x = head = hash_bucket(d, gpfn); 10.60 10.61 - SH_VVLOG("set gpfn=%08x s=%08lx bucket=%p(%p)", gpfn, s, x, x->next); 10.62 + SH_VVLOG("set gpfn=%08x s=%p bucket=%p(%p)", gpfn, s, x, x->next); 10.63 shadow_audit(d, 0); 10.64 10.65 /* 10.66 @@ -692,7 +692,7 @@ static inline void __shadow_mk_pagetable 10.67 unsigned long gpfn = pagetable_val(ed->arch.pagetable) >> PAGE_SHIFT; 10.68 unsigned long smfn = __shadow_status(d, gpfn) & PSH_pfn_mask; 10.69 10.70 - SH_VVLOG("0: __shadow_mk_pagetable(gpfn=%08lx, smfn=%08lx)", gpfn, smfn); 10.71 + SH_VVLOG("0: __shadow_mk_pagetable(gpfn=%p, smfn=%p)", gpfn, smfn); 10.72 10.73 if ( unlikely(smfn == 0) ) 10.74 smfn = shadow_l2_table(d, gpfn); 10.75 @@ -709,7 +709,7 @@ static inline void shadow_mk_pagetable(s 10.76 { 10.77 if ( unlikely(shadow_mode(ed->domain)) ) 10.78 { 10.79 - SH_VVLOG("shadow_mk_pagetable( gptbase=%08lx, mode=%d )", 10.80 + SH_VVLOG("shadow_mk_pagetable( gptbase=%p, mode=%d )", 10.81 pagetable_val(ed->arch.pagetable), 10.82 shadow_mode(ed->domain)); 10.83 10.84 @@ -718,7 +718,7 @@ static inline void shadow_mk_pagetable(s 10.85 shadow_unlock(ed->domain); 10.86 10.87 SH_VVLOG("leaving shadow_mk_pagetable:\n" 10.88 - "( gptbase=%08lx, mode=%d ) sh=%08lx", 10.89 + "( gptbase=%p, mode=%d ) sh=%p", 10.90 pagetable_val(ed->arch.pagetable), 10.91 shadow_mode(ed->domain), 10.92 pagetable_val(ed->arch.shadow_table) );
11.1 --- a/xen/include/asm-x86/x86_64/regs.h Tue Feb 08 16:44:16 2005 +0000 11.2 +++ b/xen/include/asm-x86/x86_64/regs.h Tue Feb 08 17:36:37 2005 +0000 11.3 @@ -9,18 +9,18 @@ struct xen_regs 11.4 u64 r14; 11.5 u64 r13; 11.6 u64 r12; 11.7 - union { u64 rbp; u32 ebp; } __attribute__ ((packed)); 11.8 - union { u64 rbx; u32 ebx; } __attribute__ ((packed)); 11.9 + union { u64 rbp; u64 ebp; } __attribute__ ((packed)); 11.10 + union { u64 rbx; u64 ebx; } __attribute__ ((packed)); 11.11 /* NB. Above here is C callee-saves. */ 11.12 u64 r11; 11.13 u64 r10; 11.14 u64 r9; 11.15 u64 r8; 11.16 - union { u64 rax; u32 eax; } __attribute__ ((packed)); 11.17 - union { u64 rcx; u32 ecx; } __attribute__ ((packed)); 11.18 - union { u64 rdx; u32 edx; } __attribute__ ((packed)); 11.19 - union { u64 rsi; u32 esi; } __attribute__ ((packed)); 11.20 - union { u64 rdi; u32 edi; } __attribute__ ((packed)); 11.21 + union { u64 rax; u64 eax; } __attribute__ ((packed)); 11.22 + union { u64 rcx; u64 ecx; } __attribute__ ((packed)); 11.23 + union { u64 rdx; u64 edx; } __attribute__ ((packed)); 11.24 + union { u64 rsi; u64 esi; } __attribute__ ((packed)); 11.25 + union { u64 rdi; u64 edi; } __attribute__ ((packed)); 11.26 u32 error_code; 11.27 u32 entry_vector; 11.28 union { u64 rip; u64 eip; } __attribute__ ((packed));