KVM: PPC: Use ccr field in pt_regs struct embedded in vcpu struct
When the 'regs' field was added to struct kvm_vcpu_arch, the code was changed to use several of the fields inside regs (e.g., gpr, lr, etc.) but not the ccr field, because the ccr field in struct pt_regs is 64 bits on 64-bit platforms, but the cr field in kvm_vcpu_arch is only 32 bits. This changes the code to use the regs.ccr field instead of cr, and changes the assembly code on 64-bit platforms to use 64-bit loads and stores instead of 32-bit ones. Reviewed-by: David Gibson <david@gibson.dropbear.id.au> Signed-off-by: Paul Mackerras <paulus@ozlabs.org> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
This commit is contained in:
parent
9a94d3ee2d
commit
fd0944baad
|
@ -301,12 +301,12 @@ static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num)
|
||||||
|
|
||||||
static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val)
|
static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val)
|
||||||
{
|
{
|
||||||
vcpu->arch.cr = val;
|
vcpu->arch.regs.ccr = val;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu)
|
static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
return vcpu->arch.cr;
|
return vcpu->arch.regs.ccr;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val)
|
static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val)
|
||||||
|
|
|
@ -483,7 +483,7 @@ static inline u64 sanitize_msr(u64 msr)
|
||||||
#ifdef CONFIG_PPC_TRANSACTIONAL_MEM
|
#ifdef CONFIG_PPC_TRANSACTIONAL_MEM
|
||||||
static inline void copy_from_checkpoint(struct kvm_vcpu *vcpu)
|
static inline void copy_from_checkpoint(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
vcpu->arch.cr = vcpu->arch.cr_tm;
|
vcpu->arch.regs.ccr = vcpu->arch.cr_tm;
|
||||||
vcpu->arch.regs.xer = vcpu->arch.xer_tm;
|
vcpu->arch.regs.xer = vcpu->arch.xer_tm;
|
||||||
vcpu->arch.regs.link = vcpu->arch.lr_tm;
|
vcpu->arch.regs.link = vcpu->arch.lr_tm;
|
||||||
vcpu->arch.regs.ctr = vcpu->arch.ctr_tm;
|
vcpu->arch.regs.ctr = vcpu->arch.ctr_tm;
|
||||||
|
@ -500,7 +500,7 @@ static inline void copy_from_checkpoint(struct kvm_vcpu *vcpu)
|
||||||
|
|
||||||
static inline void copy_to_checkpoint(struct kvm_vcpu *vcpu)
|
static inline void copy_to_checkpoint(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
vcpu->arch.cr_tm = vcpu->arch.cr;
|
vcpu->arch.cr_tm = vcpu->arch.regs.ccr;
|
||||||
vcpu->arch.xer_tm = vcpu->arch.regs.xer;
|
vcpu->arch.xer_tm = vcpu->arch.regs.xer;
|
||||||
vcpu->arch.lr_tm = vcpu->arch.regs.link;
|
vcpu->arch.lr_tm = vcpu->arch.regs.link;
|
||||||
vcpu->arch.ctr_tm = vcpu->arch.regs.ctr;
|
vcpu->arch.ctr_tm = vcpu->arch.regs.ctr;
|
||||||
|
|
|
@ -46,12 +46,12 @@ static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num)
|
||||||
|
|
||||||
static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val)
|
static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val)
|
||||||
{
|
{
|
||||||
vcpu->arch.cr = val;
|
vcpu->arch.regs.ccr = val;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu)
|
static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
return vcpu->arch.cr;
|
return vcpu->arch.regs.ccr;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val)
|
static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val)
|
||||||
|
|
|
@ -538,8 +538,6 @@ struct kvm_vcpu_arch {
|
||||||
ulong tar;
|
ulong tar;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
u32 cr;
|
|
||||||
|
|
||||||
#ifdef CONFIG_PPC_BOOK3S
|
#ifdef CONFIG_PPC_BOOK3S
|
||||||
ulong hflags;
|
ulong hflags;
|
||||||
ulong guest_owned_ext;
|
ulong guest_owned_ext;
|
||||||
|
|
|
@ -438,7 +438,7 @@ int main(void)
|
||||||
#ifdef CONFIG_PPC_BOOK3S
|
#ifdef CONFIG_PPC_BOOK3S
|
||||||
OFFSET(VCPU_TAR, kvm_vcpu, arch.tar);
|
OFFSET(VCPU_TAR, kvm_vcpu, arch.tar);
|
||||||
#endif
|
#endif
|
||||||
OFFSET(VCPU_CR, kvm_vcpu, arch.cr);
|
OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr);
|
||||||
OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip);
|
OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip);
|
||||||
#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE
|
#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE
|
||||||
OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr);
|
OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr);
|
||||||
|
@ -695,7 +695,7 @@ int main(void)
|
||||||
#endif /* CONFIG_PPC_BOOK3S_64 */
|
#endif /* CONFIG_PPC_BOOK3S_64 */
|
||||||
|
|
||||||
#else /* CONFIG_PPC_BOOK3S */
|
#else /* CONFIG_PPC_BOOK3S */
|
||||||
OFFSET(VCPU_CR, kvm_vcpu, arch.cr);
|
OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr);
|
||||||
OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer);
|
OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer);
|
||||||
OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link);
|
OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link);
|
||||||
OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr);
|
OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr);
|
||||||
|
|
|
@ -110,7 +110,7 @@ static inline void kvmppc_copyto_vcpu_tm(struct kvm_vcpu *vcpu)
|
||||||
vcpu->arch.ctr_tm = vcpu->arch.regs.ctr;
|
vcpu->arch.ctr_tm = vcpu->arch.regs.ctr;
|
||||||
vcpu->arch.tar_tm = vcpu->arch.tar;
|
vcpu->arch.tar_tm = vcpu->arch.tar;
|
||||||
vcpu->arch.lr_tm = vcpu->arch.regs.link;
|
vcpu->arch.lr_tm = vcpu->arch.regs.link;
|
||||||
vcpu->arch.cr_tm = vcpu->arch.cr;
|
vcpu->arch.cr_tm = vcpu->arch.regs.ccr;
|
||||||
vcpu->arch.xer_tm = vcpu->arch.regs.xer;
|
vcpu->arch.xer_tm = vcpu->arch.regs.xer;
|
||||||
vcpu->arch.vrsave_tm = vcpu->arch.vrsave;
|
vcpu->arch.vrsave_tm = vcpu->arch.vrsave;
|
||||||
}
|
}
|
||||||
|
@ -129,7 +129,7 @@ static inline void kvmppc_copyfrom_vcpu_tm(struct kvm_vcpu *vcpu)
|
||||||
vcpu->arch.regs.ctr = vcpu->arch.ctr_tm;
|
vcpu->arch.regs.ctr = vcpu->arch.ctr_tm;
|
||||||
vcpu->arch.tar = vcpu->arch.tar_tm;
|
vcpu->arch.tar = vcpu->arch.tar_tm;
|
||||||
vcpu->arch.regs.link = vcpu->arch.lr_tm;
|
vcpu->arch.regs.link = vcpu->arch.lr_tm;
|
||||||
vcpu->arch.cr = vcpu->arch.cr_tm;
|
vcpu->arch.regs.ccr = vcpu->arch.cr_tm;
|
||||||
vcpu->arch.regs.xer = vcpu->arch.xer_tm;
|
vcpu->arch.regs.xer = vcpu->arch.xer_tm;
|
||||||
vcpu->arch.vrsave = vcpu->arch.vrsave_tm;
|
vcpu->arch.vrsave = vcpu->arch.vrsave_tm;
|
||||||
}
|
}
|
||||||
|
@ -141,7 +141,7 @@ static void kvmppc_emulate_treclaim(struct kvm_vcpu *vcpu, int ra_val)
|
||||||
uint64_t texasr;
|
uint64_t texasr;
|
||||||
|
|
||||||
/* CR0 = 0 | MSR[TS] | 0 */
|
/* CR0 = 0 | MSR[TS] | 0 */
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & ~(CR0_MASK << CR0_SHIFT)) |
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & ~(CR0_MASK << CR0_SHIFT)) |
|
||||||
(((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1))
|
(((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1))
|
||||||
<< CR0_SHIFT);
|
<< CR0_SHIFT);
|
||||||
|
|
||||||
|
@ -220,7 +220,7 @@ void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val)
|
||||||
tm_abort(ra_val);
|
tm_abort(ra_val);
|
||||||
|
|
||||||
/* CR0 = 0 | MSR[TS] | 0 */
|
/* CR0 = 0 | MSR[TS] | 0 */
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & ~(CR0_MASK << CR0_SHIFT)) |
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & ~(CR0_MASK << CR0_SHIFT)) |
|
||||||
(((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1))
|
(((guest_msr & MSR_TS_MASK) >> (MSR_TS_S_LG - 1))
|
||||||
<< CR0_SHIFT);
|
<< CR0_SHIFT);
|
||||||
|
|
||||||
|
@ -494,8 +494,8 @@ int kvmppc_core_emulate_op_pr(struct kvm_run *run, struct kvm_vcpu *vcpu,
|
||||||
|
|
||||||
if (!(kvmppc_get_msr(vcpu) & MSR_PR)) {
|
if (!(kvmppc_get_msr(vcpu) & MSR_PR)) {
|
||||||
preempt_disable();
|
preempt_disable();
|
||||||
vcpu->arch.cr = (CR0_TBEGIN_FAILURE |
|
vcpu->arch.regs.ccr = (CR0_TBEGIN_FAILURE |
|
||||||
(vcpu->arch.cr & ~(CR0_MASK << CR0_SHIFT)));
|
(vcpu->arch.regs.ccr & ~(CR0_MASK << CR0_SHIFT)));
|
||||||
|
|
||||||
vcpu->arch.texasr = (TEXASR_FS | TEXASR_EXACT |
|
vcpu->arch.texasr = (TEXASR_FS | TEXASR_EXACT |
|
||||||
(((u64)(TM_CAUSE_EMULATE | TM_CAUSE_PERSISTENT))
|
(((u64)(TM_CAUSE_EMULATE | TM_CAUSE_PERSISTENT))
|
||||||
|
|
|
@ -410,8 +410,8 @@ static void kvmppc_dump_regs(struct kvm_vcpu *vcpu)
|
||||||
vcpu->arch.shregs.sprg0, vcpu->arch.shregs.sprg1);
|
vcpu->arch.shregs.sprg0, vcpu->arch.shregs.sprg1);
|
||||||
pr_err("sprg2 = %.16llx sprg3 = %.16llx\n",
|
pr_err("sprg2 = %.16llx sprg3 = %.16llx\n",
|
||||||
vcpu->arch.shregs.sprg2, vcpu->arch.shregs.sprg3);
|
vcpu->arch.shregs.sprg2, vcpu->arch.shregs.sprg3);
|
||||||
pr_err("cr = %.8x xer = %.16lx dsisr = %.8x\n",
|
pr_err("cr = %.8lx xer = %.16lx dsisr = %.8x\n",
|
||||||
vcpu->arch.cr, vcpu->arch.regs.xer, vcpu->arch.shregs.dsisr);
|
vcpu->arch.regs.ccr, vcpu->arch.regs.xer, vcpu->arch.shregs.dsisr);
|
||||||
pr_err("dar = %.16llx\n", vcpu->arch.shregs.dar);
|
pr_err("dar = %.16llx\n", vcpu->arch.shregs.dar);
|
||||||
pr_err("fault dar = %.16lx dsisr = %.8x\n",
|
pr_err("fault dar = %.16lx dsisr = %.8x\n",
|
||||||
vcpu->arch.fault_dar, vcpu->arch.fault_dsisr);
|
vcpu->arch.fault_dar, vcpu->arch.fault_dsisr);
|
||||||
|
|
|
@ -1092,7 +1092,7 @@ BEGIN_FTR_SECTION
|
||||||
END_FTR_SECTION_IFSET(CPU_FTR_HAS_PPR)
|
END_FTR_SECTION_IFSET(CPU_FTR_HAS_PPR)
|
||||||
|
|
||||||
ld r5, VCPU_LR(r4)
|
ld r5, VCPU_LR(r4)
|
||||||
lwz r6, VCPU_CR(r4)
|
ld r6, VCPU_CR(r4)
|
||||||
mtlr r5
|
mtlr r5
|
||||||
mtcr r6
|
mtcr r6
|
||||||
|
|
||||||
|
@ -1280,7 +1280,7 @@ kvmppc_interrupt_hv:
|
||||||
std r3, VCPU_GPR(R12)(r9)
|
std r3, VCPU_GPR(R12)(r9)
|
||||||
/* CR is in the high half of r12 */
|
/* CR is in the high half of r12 */
|
||||||
srdi r4, r12, 32
|
srdi r4, r12, 32
|
||||||
stw r4, VCPU_CR(r9)
|
std r4, VCPU_CR(r9)
|
||||||
BEGIN_FTR_SECTION
|
BEGIN_FTR_SECTION
|
||||||
ld r3, HSTATE_CFAR(r13)
|
ld r3, HSTATE_CFAR(r13)
|
||||||
std r3, VCPU_CFAR(r9)
|
std r3, VCPU_CFAR(r9)
|
||||||
|
|
|
@ -130,7 +130,7 @@ int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu)
|
||||||
return RESUME_GUEST;
|
return RESUME_GUEST;
|
||||||
}
|
}
|
||||||
/* Set CR0 to indicate previous transactional state */
|
/* Set CR0 to indicate previous transactional state */
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & 0x0fffffff) |
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
|
||||||
(((msr & MSR_TS_MASK) >> MSR_TS_S_LG) << 28);
|
(((msr & MSR_TS_MASK) >> MSR_TS_S_LG) << 28);
|
||||||
/* L=1 => tresume, L=0 => tsuspend */
|
/* L=1 => tresume, L=0 => tsuspend */
|
||||||
if (instr & (1 << 21)) {
|
if (instr & (1 << 21)) {
|
||||||
|
@ -174,7 +174,7 @@ int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu)
|
||||||
copy_from_checkpoint(vcpu);
|
copy_from_checkpoint(vcpu);
|
||||||
|
|
||||||
/* Set CR0 to indicate previous transactional state */
|
/* Set CR0 to indicate previous transactional state */
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & 0x0fffffff) |
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
|
||||||
(((msr & MSR_TS_MASK) >> MSR_TS_S_LG) << 28);
|
(((msr & MSR_TS_MASK) >> MSR_TS_S_LG) << 28);
|
||||||
vcpu->arch.shregs.msr &= ~MSR_TS_MASK;
|
vcpu->arch.shregs.msr &= ~MSR_TS_MASK;
|
||||||
return RESUME_GUEST;
|
return RESUME_GUEST;
|
||||||
|
@ -204,7 +204,7 @@ int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu)
|
||||||
copy_to_checkpoint(vcpu);
|
copy_to_checkpoint(vcpu);
|
||||||
|
|
||||||
/* Set CR0 to indicate previous transactional state */
|
/* Set CR0 to indicate previous transactional state */
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & 0x0fffffff) |
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
|
||||||
(((msr & MSR_TS_MASK) >> MSR_TS_S_LG) << 28);
|
(((msr & MSR_TS_MASK) >> MSR_TS_S_LG) << 28);
|
||||||
vcpu->arch.shregs.msr = msr | MSR_TS_S;
|
vcpu->arch.shregs.msr = msr | MSR_TS_S;
|
||||||
return RESUME_GUEST;
|
return RESUME_GUEST;
|
||||||
|
|
|
@ -89,7 +89,8 @@ int kvmhv_p9_tm_emulation_early(struct kvm_vcpu *vcpu)
|
||||||
if (instr & (1 << 21))
|
if (instr & (1 << 21))
|
||||||
vcpu->arch.shregs.msr = (msr & ~MSR_TS_MASK) | MSR_TS_T;
|
vcpu->arch.shregs.msr = (msr & ~MSR_TS_MASK) | MSR_TS_T;
|
||||||
/* Set CR0 to 0b0010 */
|
/* Set CR0 to 0b0010 */
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & 0x0fffffff) | 0x20000000;
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) |
|
||||||
|
0x20000000;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -105,5 +106,5 @@ void kvmhv_emulate_tm_rollback(struct kvm_vcpu *vcpu)
|
||||||
vcpu->arch.shregs.msr &= ~MSR_TS_MASK; /* go to N state */
|
vcpu->arch.shregs.msr &= ~MSR_TS_MASK; /* go to N state */
|
||||||
vcpu->arch.regs.nip = vcpu->arch.tfhar;
|
vcpu->arch.regs.nip = vcpu->arch.tfhar;
|
||||||
copy_from_checkpoint(vcpu);
|
copy_from_checkpoint(vcpu);
|
||||||
vcpu->arch.cr = (vcpu->arch.cr & 0x0fffffff) | 0xa0000000;
|
vcpu->arch.regs.ccr = (vcpu->arch.regs.ccr & 0x0fffffff) | 0xa0000000;
|
||||||
}
|
}
|
||||||
|
|
|
@ -167,7 +167,7 @@ void kvmppc_copy_to_svcpu(struct kvm_vcpu *vcpu)
|
||||||
svcpu->gpr[11] = vcpu->arch.regs.gpr[11];
|
svcpu->gpr[11] = vcpu->arch.regs.gpr[11];
|
||||||
svcpu->gpr[12] = vcpu->arch.regs.gpr[12];
|
svcpu->gpr[12] = vcpu->arch.regs.gpr[12];
|
||||||
svcpu->gpr[13] = vcpu->arch.regs.gpr[13];
|
svcpu->gpr[13] = vcpu->arch.regs.gpr[13];
|
||||||
svcpu->cr = vcpu->arch.cr;
|
svcpu->cr = vcpu->arch.regs.ccr;
|
||||||
svcpu->xer = vcpu->arch.regs.xer;
|
svcpu->xer = vcpu->arch.regs.xer;
|
||||||
svcpu->ctr = vcpu->arch.regs.ctr;
|
svcpu->ctr = vcpu->arch.regs.ctr;
|
||||||
svcpu->lr = vcpu->arch.regs.link;
|
svcpu->lr = vcpu->arch.regs.link;
|
||||||
|
@ -249,7 +249,7 @@ void kvmppc_copy_from_svcpu(struct kvm_vcpu *vcpu)
|
||||||
vcpu->arch.regs.gpr[11] = svcpu->gpr[11];
|
vcpu->arch.regs.gpr[11] = svcpu->gpr[11];
|
||||||
vcpu->arch.regs.gpr[12] = svcpu->gpr[12];
|
vcpu->arch.regs.gpr[12] = svcpu->gpr[12];
|
||||||
vcpu->arch.regs.gpr[13] = svcpu->gpr[13];
|
vcpu->arch.regs.gpr[13] = svcpu->gpr[13];
|
||||||
vcpu->arch.cr = svcpu->cr;
|
vcpu->arch.regs.ccr = svcpu->cr;
|
||||||
vcpu->arch.regs.xer = svcpu->xer;
|
vcpu->arch.regs.xer = svcpu->xer;
|
||||||
vcpu->arch.regs.ctr = svcpu->ctr;
|
vcpu->arch.regs.ctr = svcpu->ctr;
|
||||||
vcpu->arch.regs.link = svcpu->lr;
|
vcpu->arch.regs.link = svcpu->lr;
|
||||||
|
|
|
@ -182,7 +182,7 @@
|
||||||
*/
|
*/
|
||||||
PPC_LL r4, PACACURRENT(r13)
|
PPC_LL r4, PACACURRENT(r13)
|
||||||
PPC_LL r4, (THREAD + THREAD_KVM_VCPU)(r4)
|
PPC_LL r4, (THREAD + THREAD_KVM_VCPU)(r4)
|
||||||
stw r10, VCPU_CR(r4)
|
PPC_STL r10, VCPU_CR(r4)
|
||||||
PPC_STL r11, VCPU_GPR(R4)(r4)
|
PPC_STL r11, VCPU_GPR(R4)(r4)
|
||||||
PPC_STL r5, VCPU_GPR(R5)(r4)
|
PPC_STL r5, VCPU_GPR(R5)(r4)
|
||||||
PPC_STL r6, VCPU_GPR(R6)(r4)
|
PPC_STL r6, VCPU_GPR(R6)(r4)
|
||||||
|
@ -292,7 +292,7 @@ _GLOBAL(kvmppc_handler_\intno\()_\srr1)
|
||||||
PPC_STL r4, VCPU_GPR(R4)(r11)
|
PPC_STL r4, VCPU_GPR(R4)(r11)
|
||||||
PPC_LL r4, THREAD_NORMSAVE(0)(r10)
|
PPC_LL r4, THREAD_NORMSAVE(0)(r10)
|
||||||
PPC_STL r5, VCPU_GPR(R5)(r11)
|
PPC_STL r5, VCPU_GPR(R5)(r11)
|
||||||
stw r13, VCPU_CR(r11)
|
PPC_STL r13, VCPU_CR(r11)
|
||||||
mfspr r5, \srr0
|
mfspr r5, \srr0
|
||||||
PPC_STL r3, VCPU_GPR(R10)(r11)
|
PPC_STL r3, VCPU_GPR(R10)(r11)
|
||||||
PPC_LL r3, THREAD_NORMSAVE(2)(r10)
|
PPC_LL r3, THREAD_NORMSAVE(2)(r10)
|
||||||
|
@ -319,7 +319,7 @@ _GLOBAL(kvmppc_handler_\intno\()_\srr1)
|
||||||
PPC_STL r4, VCPU_GPR(R4)(r11)
|
PPC_STL r4, VCPU_GPR(R4)(r11)
|
||||||
PPC_LL r4, GPR9(r8)
|
PPC_LL r4, GPR9(r8)
|
||||||
PPC_STL r5, VCPU_GPR(R5)(r11)
|
PPC_STL r5, VCPU_GPR(R5)(r11)
|
||||||
stw r9, VCPU_CR(r11)
|
PPC_STL r9, VCPU_CR(r11)
|
||||||
mfspr r5, \srr0
|
mfspr r5, \srr0
|
||||||
PPC_STL r3, VCPU_GPR(R8)(r11)
|
PPC_STL r3, VCPU_GPR(R8)(r11)
|
||||||
PPC_LL r3, GPR10(r8)
|
PPC_LL r3, GPR10(r8)
|
||||||
|
@ -643,7 +643,7 @@ lightweight_exit:
|
||||||
PPC_LL r3, VCPU_LR(r4)
|
PPC_LL r3, VCPU_LR(r4)
|
||||||
PPC_LL r5, VCPU_XER(r4)
|
PPC_LL r5, VCPU_XER(r4)
|
||||||
PPC_LL r6, VCPU_CTR(r4)
|
PPC_LL r6, VCPU_CTR(r4)
|
||||||
lwz r7, VCPU_CR(r4)
|
PPC_LL r7, VCPU_CR(r4)
|
||||||
PPC_LL r8, VCPU_PC(r4)
|
PPC_LL r8, VCPU_PC(r4)
|
||||||
PPC_LD(r9, VCPU_SHARED_MSR, r11)
|
PPC_LD(r9, VCPU_SHARED_MSR, r11)
|
||||||
PPC_LL r0, VCPU_GPR(R0)(r4)
|
PPC_LL r0, VCPU_GPR(R0)(r4)
|
||||||
|
|
|
@ -117,7 +117,6 @@ int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu)
|
||||||
|
|
||||||
emulated = EMULATE_FAIL;
|
emulated = EMULATE_FAIL;
|
||||||
vcpu->arch.regs.msr = vcpu->arch.shared->msr;
|
vcpu->arch.regs.msr = vcpu->arch.shared->msr;
|
||||||
vcpu->arch.regs.ccr = vcpu->arch.cr;
|
|
||||||
if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) {
|
if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) {
|
||||||
int type = op.type & INSTR_TYPE_MASK;
|
int type = op.type & INSTR_TYPE_MASK;
|
||||||
int size = GETSIZE(op.type);
|
int size = GETSIZE(op.type);
|
||||||
|
|
Loading…
Reference in New Issue