KVM: MMU: rename 'page' and 'shadow_page' to 'sp'
Rename 'page' and 'shadow_page' to 'sp' to better fit the context Signed-off-by: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com> Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
This commit is contained in:
parent
2d5b5a6655
commit
ac3cd03cca
|
@ -253,7 +253,7 @@ err:
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
|
static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
|
||||||
u64 *spte, const void *pte)
|
u64 *spte, const void *pte)
|
||||||
{
|
{
|
||||||
pt_element_t gpte;
|
pt_element_t gpte;
|
||||||
|
@ -264,7 +264,7 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
|
||||||
gpte = *(const pt_element_t *)pte;
|
gpte = *(const pt_element_t *)pte;
|
||||||
if (~gpte & (PT_PRESENT_MASK | PT_ACCESSED_MASK)) {
|
if (~gpte & (PT_PRESENT_MASK | PT_ACCESSED_MASK)) {
|
||||||
if (!is_present_gpte(gpte)) {
|
if (!is_present_gpte(gpte)) {
|
||||||
if (page->unsync)
|
if (sp->unsync)
|
||||||
new_spte = shadow_trap_nonpresent_pte;
|
new_spte = shadow_trap_nonpresent_pte;
|
||||||
else
|
else
|
||||||
new_spte = shadow_notrap_nonpresent_pte;
|
new_spte = shadow_notrap_nonpresent_pte;
|
||||||
|
@ -273,7 +273,7 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
pgprintk("%s: gpte %llx spte %p\n", __func__, (u64)gpte, spte);
|
pgprintk("%s: gpte %llx spte %p\n", __func__, (u64)gpte, spte);
|
||||||
pte_access = page->role.access & FNAME(gpte_access)(vcpu, gpte);
|
pte_access = sp->role.access & FNAME(gpte_access)(vcpu, gpte);
|
||||||
if (gpte_to_gfn(gpte) != vcpu->arch.update_pte.gfn)
|
if (gpte_to_gfn(gpte) != vcpu->arch.update_pte.gfn)
|
||||||
return;
|
return;
|
||||||
pfn = vcpu->arch.update_pte.pfn;
|
pfn = vcpu->arch.update_pte.pfn;
|
||||||
|
@ -286,7 +286,7 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
|
||||||
* we call mmu_set_spte() with reset_host_protection = true beacuse that
|
* we call mmu_set_spte() with reset_host_protection = true beacuse that
|
||||||
* vcpu->arch.update_pte.pfn was fetched from get_user_pages(write = 1).
|
* vcpu->arch.update_pte.pfn was fetched from get_user_pages(write = 1).
|
||||||
*/
|
*/
|
||||||
mmu_set_spte(vcpu, spte, page->role.access, pte_access, 0, 0,
|
mmu_set_spte(vcpu, spte, sp->role.access, pte_access, 0, 0,
|
||||||
gpte & PT_DIRTY_MASK, NULL, PT_PAGE_TABLE_LEVEL,
|
gpte & PT_DIRTY_MASK, NULL, PT_PAGE_TABLE_LEVEL,
|
||||||
gpte_to_gfn(gpte), pfn, true, true);
|
gpte_to_gfn(gpte), pfn, true, true);
|
||||||
}
|
}
|
||||||
|
@ -300,7 +300,7 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
|
||||||
int *ptwrite, pfn_t pfn)
|
int *ptwrite, pfn_t pfn)
|
||||||
{
|
{
|
||||||
unsigned access = gw->pt_access;
|
unsigned access = gw->pt_access;
|
||||||
struct kvm_mmu_page *shadow_page;
|
struct kvm_mmu_page *sp;
|
||||||
u64 spte, *sptep = NULL;
|
u64 spte, *sptep = NULL;
|
||||||
int direct;
|
int direct;
|
||||||
gfn_t table_gfn;
|
gfn_t table_gfn;
|
||||||
|
@ -341,9 +341,9 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
|
||||||
access &= ~ACC_WRITE_MASK;
|
access &= ~ACC_WRITE_MASK;
|
||||||
/*
|
/*
|
||||||
* It is a large guest pages backed by small host pages,
|
* It is a large guest pages backed by small host pages,
|
||||||
* So we set @direct(@shadow_page->role.direct)=1, and
|
* So we set @direct(@sp->role.direct)=1, and set
|
||||||
* set @table_gfn(@shadow_page->gfn)=the base page frame
|
* @table_gfn(@sp->gfn)=the base page frame for linear
|
||||||
* for linear translations.
|
* translations.
|
||||||
*/
|
*/
|
||||||
table_gfn = gw->gfn & ~(KVM_PAGES_PER_HPAGE(level) - 1);
|
table_gfn = gw->gfn & ~(KVM_PAGES_PER_HPAGE(level) - 1);
|
||||||
access &= gw->pte_access;
|
access &= gw->pte_access;
|
||||||
|
@ -351,21 +351,21 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
|
||||||
direct = 0;
|
direct = 0;
|
||||||
table_gfn = gw->table_gfn[level - 2];
|
table_gfn = gw->table_gfn[level - 2];
|
||||||
}
|
}
|
||||||
shadow_page = kvm_mmu_get_page(vcpu, table_gfn, addr, level-1,
|
sp = kvm_mmu_get_page(vcpu, table_gfn, addr, level-1,
|
||||||
direct, access, sptep);
|
direct, access, sptep);
|
||||||
if (!direct) {
|
if (!direct) {
|
||||||
r = kvm_read_guest_atomic(vcpu->kvm,
|
r = kvm_read_guest_atomic(vcpu->kvm,
|
||||||
gw->pte_gpa[level - 2],
|
gw->pte_gpa[level - 2],
|
||||||
&curr_pte, sizeof(curr_pte));
|
&curr_pte, sizeof(curr_pte));
|
||||||
if (r || curr_pte != gw->ptes[level - 2]) {
|
if (r || curr_pte != gw->ptes[level - 2]) {
|
||||||
kvm_mmu_put_page(shadow_page, sptep);
|
kvm_mmu_put_page(sp, sptep);
|
||||||
kvm_release_pfn_clean(pfn);
|
kvm_release_pfn_clean(pfn);
|
||||||
sptep = NULL;
|
sptep = NULL;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
spte = __pa(shadow_page->spt)
|
spte = __pa(sp->spt)
|
||||||
| PT_PRESENT_MASK | PT_ACCESSED_MASK
|
| PT_PRESENT_MASK | PT_ACCESSED_MASK
|
||||||
| PT_WRITABLE_MASK | PT_USER_MASK;
|
| PT_WRITABLE_MASK | PT_USER_MASK;
|
||||||
*sptep = spte;
|
*sptep = spte;
|
||||||
|
|
Loading…
Reference in New Issue