Sign Up
Log In
Log In
or
Sign Up
Places
All Projects
Status Monitor
Collapse sidebar
SUSE:SLE-12-SP1:GA
xen.10697
5b72fbbf-3-x86-infrastructure-to-force-PV-guest...
Overview
Repositories
Revisions
Requests
Users
Attributes
Meta
File 5b72fbbf-3-x86-infrastructure-to-force-PV-guest-into-shadow-mode.patch of Package xen.10697
# Commit b76ec3946bf6caca2c3950b857c008bc8db6723f # Date 2018-08-14 16:56:47 +0100 # Author Juergen Gross <jgross@suse.com> # Committer Andrew Cooper <andrew.cooper3@citrix.com> x86/shadow: Infrastructure to force a PV guest into shadow mode To mitigate L1TF, we cannot alter an architecturally-legitimate PTE a PV guest chooses to write, but we can force the PV domain into shadow mode so Xen controls the PTEs which are reachable by the CPU pagewalk. Introduce new shadow mode, PG_SH_forced, and a tasklet to perform the transition. Later patches will introduce the logic to enable this mode at the appropriate time. To simplify vcpu cleanup, make tasklet_kill() idempotent with respect to tasklet_init(), which involves adding a helper to check for an uninitialised list head. This is part of XSA-273 / CVE-2018-3620. Signed-off-by: Juergen Gross <jgross@suse.com> Signed-off-by: Andrew Cooper <andrew.cooper3@citrix.com> Reviewed-by: Tim Deegan <tim@xen.org> Reviewed-by: Jan Beulich <jbeulich@suse.com> --- a/xen/arch/x86/domain.c +++ b/xen/arch/x86/domain.c @@ -64,6 +64,7 @@ #include <xen/iommu.h> #include <compat/vcpu.h> #include <asm/spec_ctrl.h> +#include <asm/shadow.h> static __read_mostly enum { PCID_OFF, @@ -625,6 +626,8 @@ int arch_domain_create(struct domain *d, rc = 0; else { + pv_l1tf_domain_init(d); + d->arch.pv_domain.gdt_ldt_l1tab = alloc_xenheap_pages(0, MEMF_node(domain_to_node(d))); if ( !d->arch.pv_domain.gdt_ldt_l1tab ) @@ -722,7 +725,10 @@ int arch_domain_create(struct domain *d, paging_final_teardown(d); free_perdomain_mappings(d); if ( is_pv_domain(d) ) + { + pv_l1tf_domain_destroy(d); free_xenheap_page(d->arch.pv_domain.gdt_ldt_l1tab); + } return rc; } @@ -730,6 +736,8 @@ void arch_domain_destroy(struct domain * { if ( has_hvm_container_domain(d) ) hvm_domain_destroy(d); + else + pv_l1tf_domain_destroy(d); xfree(d->arch.e820); xfree(d->arch.cpuids); --- a/xen/arch/x86/mm/paging.c +++ b/xen/arch/x86/mm/paging.c @@ -848,6 +848,8 @@ void paging_dump_domain_info(struct doma printk(" paging assistance: "); if ( paging_mode_shadow(d) ) printk("shadow "); + if ( paging_mode_sh_forced(d) ) + printk("forced "); if ( paging_mode_hap(d) ) printk("hap "); if ( paging_mode_refcounts(d) ) --- a/xen/arch/x86/mm/shadow/common.c +++ b/xen/arch/x86/mm/shadow/common.c @@ -2923,6 +2923,15 @@ static void sh_new_mode(struct domain *d ASSERT(paging_locked_by_me(d)); ASSERT(d != current->domain); + /* + * If PG_SH_forced has previously been activated because of writing an + * L1TF-vulnerable PTE, it must remain active for the remaining lifetime + * of the domain, even if the logdirty mode needs to be controlled for + * migration purposes. + */ + if ( paging_mode_sh_forced(d) ) + new_mode |= PG_SH_forced | PG_SH_enable; + d->arch.paging.mode = new_mode; for_each_vcpu(d, v) sh_update_paging_modes(v); @@ -3807,6 +3816,29 @@ void shadow_audit_tables(struct vcpu *v) #endif /* Shadow audit */ +void pv_l1tf_tasklet(unsigned long data) +{ + struct domain *d = (void *)data; + + domain_pause(d); + paging_lock(d); + + if ( !paging_mode_sh_forced(d) && !d->is_dying ) + { + int ret = shadow_one_bit_enable(d, PG_SH_forced); + + if ( ret ) + { + printk(XENLOG_G_ERR "d%d Failed to enable PG_SH_forced: %d\n", + d->domain_id, ret); + domain_crash(d); + } + } + + paging_unlock(d); + domain_unpause(d); +} + /* * Local variables: * mode: C --- a/xen/common/tasklet.c +++ b/xen/common/tasklet.c @@ -153,6 +153,10 @@ void tasklet_kill(struct tasklet *t) spin_lock_irqsave(&tasklet_lock, flags); + /* Cope with uninitialised tasklets. */ + if ( list_head_is_null(&t->list) ) + goto unlock; + if ( !list_empty(&t->list) ) { BUG_ON(t->is_dead || t->is_running || (t->scheduled_on < 0)); @@ -169,6 +173,7 @@ void tasklet_kill(struct tasklet *t) spin_lock_irqsave(&tasklet_lock, flags); } + unlock: spin_unlock_irqrestore(&tasklet_lock, flags); } --- a/xen/include/asm-x86/domain.h +++ b/xen/include/asm-x86/domain.h @@ -117,6 +117,9 @@ struct shadow_domain { /* Has this domain ever used HVMOP_pagetable_dying? */ bool_t pagetable_dying_op; + + /* PV L1 Terminal Fault mitigation. */ + struct tasklet pv_l1tf_tasklet; }; struct shadow_vcpu { @@ -255,6 +258,8 @@ struct pv_domain bool_t xpti; /* Use PCID feature? */ bool_t pcid; + /* Mitigate L1TF with shadow/crashing? */ + bool_t check_l1tf; /* map_domain_page() mapping cache. */ struct mapcache_domain mapcache; --- a/xen/include/asm-x86/paging.h +++ b/xen/include/asm-x86/paging.h @@ -38,8 +38,10 @@ #define PG_SH_shift 20 #define PG_HAP_shift 21 +#define PG_SHF_shift 22 /* We're in one of the shadow modes */ #define PG_SH_enable (1U << PG_SH_shift) +#define PG_SH_forced (1U << PG_SHF_shift) #define PG_HAP_enable (1U << PG_HAP_shift) /* common paging mode bits */ @@ -56,6 +58,7 @@ #define paging_mode_enabled(_d) ((_d)->arch.paging.mode) #define paging_mode_shadow(_d) ((_d)->arch.paging.mode & PG_SH_enable) +#define paging_mode_sh_forced(_d) ((_d)->arch.paging.mode & PG_SH_forced) #define paging_mode_hap(_d) ((_d)->arch.paging.mode & PG_HAP_enable) #define paging_mode_refcounts(_d) ((_d)->arch.paging.mode & PG_refcounts) --- a/xen/include/asm-x86/shadow.h +++ b/xen/include/asm-x86/shadow.h @@ -30,6 +30,7 @@ #include <asm/flushtlb.h> #include <asm/paging.h> #include <asm/p2m.h> +#include <asm/spec_ctrl.h> /***************************************************************************** * Macros to tell which shadow paging mode a domain is in*/ @@ -107,6 +108,33 @@ shadow_drop_references(struct domain *d, sh_remove_all_mappings(d->vcpu[0], _mfn(page_to_mfn(p))); } +/* + * Mitigations for L1TF / CVE-2018-3620 for PV guests. + * + * We cannot alter an architecturally-legitimate PTE which a PV guest has + * chosen to write, as traditional paged-out metadata is L1TF-vulnerable. + * What we can do is force a PV guest which writes a vulnerable PTE into + * shadow mode, so Xen controls the pagetables which are reachable by the CPU + * pagewalk. + */ + +void pv_l1tf_tasklet(unsigned long data); + +static inline void pv_l1tf_domain_init(struct domain *d) +{ + d->arch.pv_domain.check_l1tf = + opt_pv_l1tf & (d->domain_id == 0 /* ->is_privileged not usable yet */ + ? OPT_PV_L1TF_DOM0 : OPT_PV_L1TF_DOMU); + + tasklet_init(&d->arch.paging.shadow.pv_l1tf_tasklet, + pv_l1tf_tasklet, (unsigned long)d); +} + +static inline void pv_l1tf_domain_destroy(struct domain *d) +{ + tasklet_kill(&d->arch.paging.shadow.pv_l1tf_tasklet); +} + /* Remove all shadows of the guest mfn. */ void sh_remove_shadows(struct vcpu *v, mfn_t gmfn, int fast, int all); static inline void shadow_remove_all_shadows(struct vcpu *v, mfn_t gmfn) --- a/xen/include/xen/list.h +++ b/xen/include/xen/list.h @@ -51,6 +51,11 @@ static inline void INIT_LIST_HEAD(struct list->prev = list; } +static inline bool_t list_head_is_null(const struct list_head *list) +{ + return !list->next && !list->prev; +} + /* * Insert a new entry between two known consecutive entries. *
Locations
Projects
Search
Status Monitor
Help
OpenBuildService.org
Documentation
API Documentation
Code of Conduct
Contact
Support
@OBShq
Terms
openSUSE Build Service is sponsored by
The Open Build Service is an
openSUSE project
.
Sign Up
Log In
Places
Places
All Projects
Status Monitor