Sign Up
Log In
Log In
or
Sign Up
Places
All Projects
Status Monitor
Collapse sidebar
SUSE:SLE-12:Update
xen.10697
5bf4135a-x86-shrink-page_info-shadow_flags.patch
Overview
Repositories
Revisions
Requests
Users
Attributes
Meta
File 5bf4135a-x86-shrink-page_info-shadow_flags.patch of Package xen.10697
# Commit 789589968ed90e82a832dbc60e958c76b787be7e # Date 2018-11-20 14:59:54 +0100 # Author Jan Beulich <jbeulich@suse.com> # Committer Jan Beulich <jbeulich@suse.com> x86/shadow: shrink struct page_info's shadow_flags to 16 bits This is to avoid it overlapping the linear_pt_count field needed for PV domains. Introduce a separate, HVM-only pagetable_dying field to replace the sole one left in the upper 16 bits. Note that the accesses to ->shadow_flags in shadow_{pro,de}mote() get switched to non-atomic, non-bitops operations, as {test,set,clear}_bit() are not allowed on uint16_t fields and hence their use would have required ugly casts. This is fine because all updates of the field ought to occur with the paging lock held, and other updates of it use |= and &= as well (i.e. using atomic operations here didn't really guard against potentially racing updates elsewhere). This is part of XSA-280. Reported-by: Prgmr.com Security <security@prgmr.com> Signed-off-by: Jan Beulich <jbeulich@suse.com> Reviewed-by: Tim Deegan <tim@xen.org> --- a/xen/arch/x86/mm/shadow/common.c +++ b/xen/arch/x86/mm/shadow/common.c @@ -1024,10 +1024,14 @@ void shadow_promote(struct vcpu *v, mfn_ /* Is the page already shadowed? */ if ( !test_and_set_bit(_PGC_page_table, &page->count_info) ) + { page->shadow_flags = 0; + if ( is_hvm_vcpu(v) ) + page->pagetable_dying = 0; + } - ASSERT(!test_bit(type, &page->shadow_flags)); - set_bit(type, &page->shadow_flags); + ASSERT(!(page->shadow_flags & (1u << type))); + page->shadow_flags |= 1u << type; TRACE_SHADOW_PATH_FLAG(TRCE_SFLAG_PROMOTE); } @@ -1036,9 +1040,9 @@ void shadow_demote(struct vcpu *v, mfn_t struct page_info *page = mfn_to_page(gmfn); ASSERT(test_bit(_PGC_page_table, &page->count_info)); - ASSERT(test_bit(type, &page->shadow_flags)); + ASSERT(page->shadow_flags & (1u << type)); - clear_bit(type, &page->shadow_flags); + page->shadow_flags &= ~(1u << type); if ( (page->shadow_flags & SHF_page_type_mask) == 0 ) { @@ -2657,7 +2661,7 @@ void sh_remove_shadows(struct vcpu *v, m if ( !fast && all && (pg->count_info & PGC_page_table) ) { SHADOW_ERROR("can't find all shadows of mfn %05lx " - "(shadow_flags=%08x)\n", + "(shadow_flags=%04x)\n", mfn_x(gmfn), pg->shadow_flags); domain_crash(v->domain); } --- a/xen/arch/x86/mm/shadow/multi.c +++ b/xen/arch/x86/mm/shadow/multi.c @@ -3306,8 +3306,8 @@ static int sh_page_fault(struct vcpu *v, /* Unshadow if we are writing to a toplevel pagetable that is * flagged as a dying process, and that is not currently used. */ - if ( sh_mfn_is_a_page_table(gmfn) - && (mfn_to_page(gmfn)->shadow_flags & SHF_pagetable_dying) ) + if ( sh_mfn_is_a_page_table(gmfn) && is_hvm_domain(d) && + mfn_to_page(gmfn)->pagetable_dying ) { int used = 0; struct vcpu *tmp; @@ -4229,8 +4229,9 @@ int sh_rm_write_access_from_sl1p(struct ASSERT(mfn_valid(smfn)); /* Remember if we've been told that this process is being torn down */ - v->arch.paging.shadow.pagetable_dying - = !!(mfn_to_page(gmfn)->shadow_flags & SHF_pagetable_dying); + if ( is_hvm_vcpu(v) ) + v->arch.paging.shadow.pagetable_dying + = mfn_to_page(gmfn)->pagetable_dying; sp = mfn_to_page(smfn); @@ -4544,10 +4545,10 @@ static void sh_pagetable_dying(struct vc smfn = shadow_hash_lookup(v, mfn_x(gmfn), SH_type_l2_pae_shadow); } - if ( mfn_valid(smfn) ) + if ( mfn_valid(smfn) && is_hvm_vcpu(v) ) { gmfn = _mfn(mfn_to_page(smfn)->v.sh.back); - mfn_to_page(gmfn)->shadow_flags |= SHF_pagetable_dying; + mfn_to_page(gmfn)->pagetable_dying = 1; shadow_unhook_mappings(v, smfn, 1/* user pages only */); flush = 1; } @@ -4583,9 +4584,9 @@ static void sh_pagetable_dying(struct vc smfn = shadow_hash_lookup(v, mfn_x(gmfn), SH_type_l4_64_shadow); #endif - if ( mfn_valid(smfn) ) + if ( mfn_valid(smfn) && is_hvm_vcpu(v) ) { - mfn_to_page(gmfn)->shadow_flags |= SHF_pagetable_dying; + mfn_to_page(gmfn)->pagetable_dying = 1; shadow_unhook_mappings(v, smfn, 1/* user pages only */); /* Now flush the TLB: we removed toplevel mappings. */ flush_tlb_mask(v->domain->domain_dirty_cpumask); --- a/xen/arch/x86/mm/shadow/private.h +++ b/xen/arch/x86/mm/shadow/private.h @@ -280,8 +280,6 @@ static inline int sh_type_has_up_pointer #endif /* (SHADOW_OPTIMIZATIONS & SHOPT_OUT_OF_SYNC) */ -#define SHF_pagetable_dying (1u<<31) - static inline int sh_page_has_multiple_shadows(struct page_info *pg) { u32 shadows; --- a/xen/include/asm-x86/mm.h +++ b/xen/include/asm-x86/mm.h @@ -166,8 +166,15 @@ struct page_info * Guest pages with a shadow. This does not conflict with * tlbflush_timestamp since page table pages are explicitly not * tracked for TLB-flush avoidance when a guest runs in shadow mode. + * + * pagetable_dying is used for HVM domains only. The layout here has + * to avoid re-use of the space used by linear_pt_count, which (only) + * PV guests use. */ - u32 shadow_flags; + struct { + uint16_t shadow_flags; + bool_t pagetable_dying; + }; /* When in use as a shadow, next shadow in this hash chain. */ __pdx_t next_shadow;
Locations
Projects
Search
Status Monitor
Help
OpenBuildService.org
Documentation
API Documentation
Code of Conduct
Contact
Support
@OBShq
Terms
openSUSE Build Service is sponsored by
The Open Build Service is an
openSUSE project
.
Sign Up
Log In
Places
Places
All Projects
Status Monitor