From c05b1a4186037d67b25767ee654560505232bfd7 Mon Sep 17 00:00:00 2001 From: Tim Deegan Date: Thu, 10 Nov 2011 11:12:35 +0000 Subject: [PATCH] x86/mm: Refactor p2m get_entry accessor Move the main query accessor to the p2m outside of an inline and into the p2m code itself. This will allow for p2m internal locking to be added to the accessor later. Signed-off-by: Andres Lagar-Cavilla Acked-by: Tim Deegan Committed-by: Tim Deegan --- xen/arch/x86/mm/p2m.c | 38 +++++++++++++++++++++++++++++++++++++ xen/include/asm-x86/p2m.h | 40 ++------------------------------------- 2 files changed, 40 insertions(+), 38 deletions(-) diff --git a/xen/arch/x86/mm/p2m.c b/xen/arch/x86/mm/p2m.c index 93ee1a811f..63d0248d06 100644 --- a/xen/arch/x86/mm/p2m.c +++ b/xen/arch/x86/mm/p2m.c @@ -144,6 +144,44 @@ void p2m_change_entry_type_global(struct domain *d, p2m_unlock(p2m); } +mfn_t gfn_to_mfn_type_p2m(struct p2m_domain *p2m, unsigned long gfn, + p2m_type_t *t, p2m_access_t *a, p2m_query_t q, + unsigned int *page_order) +{ + mfn_t mfn; + + if ( !p2m || !paging_mode_translate(p2m->domain) ) + { + /* Not necessarily true, but for non-translated guests, we claim + * it's the most generic kind of memory */ + *t = p2m_ram_rw; + return _mfn(gfn); + } + + mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order); + +#ifdef __x86_64__ + if ( q == p2m_unshare && p2m_is_shared(*t) ) + { + ASSERT(!p2m_is_nestedp2m(p2m)); + mem_sharing_unshare_page(p2m->domain, gfn, 0); + mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order); + } +#endif + +#ifdef __x86_64__ + if (unlikely((p2m_is_broken(*t)))) + { + /* Return invalid_mfn to avoid caller's access */ + mfn = _mfn(INVALID_MFN); + if (q == p2m_guest) + domain_crash(p2m->domain); + } +#endif + + return mfn; +} + int set_p2m_entry(struct p2m_domain *p2m, unsigned long gfn, mfn_t mfn, unsigned int page_order, p2m_type_t p2mt, p2m_access_t p2ma) { diff --git a/xen/include/asm-x86/p2m.h b/xen/include/asm-x86/p2m.h index 140c537480..3938626040 100644 --- a/xen/include/asm-x86/p2m.h +++ b/xen/include/asm-x86/p2m.h @@ -313,45 +313,9 @@ struct p2m_domain *p2m_get_p2m(struct vcpu *v); * If the lookup succeeds, the return value is != INVALID_MFN and * *page_order is filled in with the order of the superpage (if any) that * the entry was found in. */ -static inline mfn_t -gfn_to_mfn_type_p2m(struct p2m_domain *p2m, unsigned long gfn, +mfn_t gfn_to_mfn_type_p2m(struct p2m_domain *p2m, unsigned long gfn, p2m_type_t *t, p2m_access_t *a, p2m_query_t q, - unsigned int *page_order) -{ - mfn_t mfn; - - if ( !p2m || !paging_mode_translate(p2m->domain) ) - { - /* Not necessarily true, but for non-translated guests, we claim - * it's the most generic kind of memory */ - *t = p2m_ram_rw; - return _mfn(gfn); - } - - mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order); - -#ifdef __x86_64__ - if ( q == p2m_unshare && p2m_is_shared(*t) ) - { - ASSERT(!p2m_is_nestedp2m(p2m)); - mem_sharing_unshare_page(p2m->domain, gfn, 0); - mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order); - } -#endif - -#ifdef __x86_64__ - if (unlikely((p2m_is_broken(*t)))) - { - /* Return invalid_mfn to avoid caller's access */ - mfn = _mfn(INVALID_MFN); - if (q == p2m_guest) - domain_crash(p2m->domain); - } -#endif - - return mfn; -} - + unsigned int *page_order); /* General conversion function from gfn to mfn */ static inline mfn_t gfn_to_mfn_type(struct domain *d, -- 2.30.2