mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-30 14:46:47 +07:00
0f0412c1a7
kvm_get_exit_data() cannot return a NULL pointer. Signed-off-by: Roel Kluin <roel.kluin@gmail.com> Signed-off-by: Avi Kivity <avi@redhat.com>
675 lines
16 KiB
C
675 lines
16 KiB
C
/*
|
|
* PAL/SAL call delegation
|
|
*
|
|
* Copyright (c) 2004 Li Susie <susie.li@intel.com>
|
|
* Copyright (c) 2005 Yu Ke <ke.yu@intel.com>
|
|
* Copyright (c) 2007 Xiantao Zhang <xiantao.zhang@intel.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program; if not, write to the Free Software Foundation, Inc., 59 Temple
|
|
* Place - Suite 330, Boston, MA 02111-1307 USA.
|
|
*/
|
|
|
|
#include <linux/kvm_host.h>
|
|
#include <linux/smp.h>
|
|
#include <asm/sn/addrs.h>
|
|
#include <asm/sn/clksupport.h>
|
|
#include <asm/sn/shub_mmr.h>
|
|
|
|
#include "vti.h"
|
|
#include "misc.h"
|
|
|
|
#include <asm/pal.h>
|
|
#include <asm/sal.h>
|
|
#include <asm/tlb.h>
|
|
|
|
/*
|
|
* Handy macros to make sure that the PAL return values start out
|
|
* as something meaningful.
|
|
*/
|
|
#define INIT_PAL_STATUS_UNIMPLEMENTED(x) \
|
|
{ \
|
|
x.status = PAL_STATUS_UNIMPLEMENTED; \
|
|
x.v0 = 0; \
|
|
x.v1 = 0; \
|
|
x.v2 = 0; \
|
|
}
|
|
|
|
#define INIT_PAL_STATUS_SUCCESS(x) \
|
|
{ \
|
|
x.status = PAL_STATUS_SUCCESS; \
|
|
x.v0 = 0; \
|
|
x.v1 = 0; \
|
|
x.v2 = 0; \
|
|
}
|
|
|
|
static void kvm_get_pal_call_data(struct kvm_vcpu *vcpu,
|
|
u64 *gr28, u64 *gr29, u64 *gr30, u64 *gr31) {
|
|
struct exit_ctl_data *p;
|
|
|
|
if (vcpu) {
|
|
p = &vcpu->arch.exit_data;
|
|
if (p->exit_reason == EXIT_REASON_PAL_CALL) {
|
|
*gr28 = p->u.pal_data.gr28;
|
|
*gr29 = p->u.pal_data.gr29;
|
|
*gr30 = p->u.pal_data.gr30;
|
|
*gr31 = p->u.pal_data.gr31;
|
|
return ;
|
|
}
|
|
}
|
|
printk(KERN_DEBUG"Failed to get vcpu pal data!!!\n");
|
|
}
|
|
|
|
static void set_pal_result(struct kvm_vcpu *vcpu,
|
|
struct ia64_pal_retval result) {
|
|
|
|
struct exit_ctl_data *p;
|
|
|
|
p = kvm_get_exit_data(vcpu);
|
|
if (p->exit_reason == EXIT_REASON_PAL_CALL) {
|
|
p->u.pal_data.ret = result;
|
|
return ;
|
|
}
|
|
INIT_PAL_STATUS_UNIMPLEMENTED(p->u.pal_data.ret);
|
|
}
|
|
|
|
static void set_sal_result(struct kvm_vcpu *vcpu,
|
|
struct sal_ret_values result) {
|
|
struct exit_ctl_data *p;
|
|
|
|
p = kvm_get_exit_data(vcpu);
|
|
if (p->exit_reason == EXIT_REASON_SAL_CALL) {
|
|
p->u.sal_data.ret = result;
|
|
return ;
|
|
}
|
|
printk(KERN_WARNING"Failed to set sal result!!\n");
|
|
}
|
|
|
|
struct cache_flush_args {
|
|
u64 cache_type;
|
|
u64 operation;
|
|
u64 progress;
|
|
long status;
|
|
};
|
|
|
|
cpumask_t cpu_cache_coherent_map;
|
|
|
|
static void remote_pal_cache_flush(void *data)
|
|
{
|
|
struct cache_flush_args *args = data;
|
|
long status;
|
|
u64 progress = args->progress;
|
|
|
|
status = ia64_pal_cache_flush(args->cache_type, args->operation,
|
|
&progress, NULL);
|
|
if (status != 0)
|
|
args->status = status;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_cache_flush(struct kvm_vcpu *vcpu)
|
|
{
|
|
u64 gr28, gr29, gr30, gr31;
|
|
struct ia64_pal_retval result = {0, 0, 0, 0};
|
|
struct cache_flush_args args = {0, 0, 0, 0};
|
|
long psr;
|
|
|
|
gr28 = gr29 = gr30 = gr31 = 0;
|
|
kvm_get_pal_call_data(vcpu, &gr28, &gr29, &gr30, &gr31);
|
|
|
|
if (gr31 != 0)
|
|
printk(KERN_ERR"vcpu:%p called cache_flush error!\n", vcpu);
|
|
|
|
/* Always call Host Pal in int=1 */
|
|
gr30 &= ~PAL_CACHE_FLUSH_CHK_INTRS;
|
|
args.cache_type = gr29;
|
|
args.operation = gr30;
|
|
smp_call_function(remote_pal_cache_flush,
|
|
(void *)&args, 1);
|
|
if (args.status != 0)
|
|
printk(KERN_ERR"pal_cache_flush error!,"
|
|
"status:0x%lx\n", args.status);
|
|
/*
|
|
* Call Host PAL cache flush
|
|
* Clear psr.ic when call PAL_CACHE_FLUSH
|
|
*/
|
|
local_irq_save(psr);
|
|
result.status = ia64_pal_cache_flush(gr29, gr30, &result.v1,
|
|
&result.v0);
|
|
local_irq_restore(psr);
|
|
if (result.status != 0)
|
|
printk(KERN_ERR"vcpu:%p crashed due to cache_flush err:%ld"
|
|
"in1:%lx,in2:%lx\n",
|
|
vcpu, result.status, gr29, gr30);
|
|
|
|
#if 0
|
|
if (gr29 == PAL_CACHE_TYPE_COHERENT) {
|
|
cpus_setall(vcpu->arch.cache_coherent_map);
|
|
cpu_clear(vcpu->cpu, vcpu->arch.cache_coherent_map);
|
|
cpus_setall(cpu_cache_coherent_map);
|
|
cpu_clear(vcpu->cpu, cpu_cache_coherent_map);
|
|
}
|
|
#endif
|
|
return result;
|
|
}
|
|
|
|
struct ia64_pal_retval pal_cache_summary(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
struct ia64_pal_retval result;
|
|
|
|
PAL_CALL(result, PAL_CACHE_SUMMARY, 0, 0, 0);
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_freq_base(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
struct ia64_pal_retval result;
|
|
|
|
PAL_CALL(result, PAL_FREQ_BASE, 0, 0, 0);
|
|
|
|
/*
|
|
* PAL_FREQ_BASE may not be implemented in some platforms,
|
|
* call SAL instead.
|
|
*/
|
|
if (result.v0 == 0) {
|
|
result.status = ia64_sal_freq_base(SAL_FREQ_BASE_PLATFORM,
|
|
&result.v0,
|
|
&result.v1);
|
|
result.v2 = 0;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/*
|
|
* On the SGI SN2, the ITC isn't stable. Emulation backed by the SN2
|
|
* RTC is used instead. This function patches the ratios from SAL
|
|
* to match the RTC before providing them to the guest.
|
|
*/
|
|
static void sn2_patch_itc_freq_ratios(struct ia64_pal_retval *result)
|
|
{
|
|
struct pal_freq_ratio *ratio;
|
|
unsigned long sal_freq, sal_drift, factor;
|
|
|
|
result->status = ia64_sal_freq_base(SAL_FREQ_BASE_PLATFORM,
|
|
&sal_freq, &sal_drift);
|
|
ratio = (struct pal_freq_ratio *)&result->v2;
|
|
factor = ((sal_freq * 3) + (sn_rtc_cycles_per_second / 2)) /
|
|
sn_rtc_cycles_per_second;
|
|
|
|
ratio->num = 3;
|
|
ratio->den = factor;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_freq_ratios(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct ia64_pal_retval result;
|
|
|
|
PAL_CALL(result, PAL_FREQ_RATIOS, 0, 0, 0);
|
|
|
|
if (vcpu->kvm->arch.is_sn2)
|
|
sn2_patch_itc_freq_ratios(&result);
|
|
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_logical_to_physica(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct ia64_pal_retval result;
|
|
|
|
INIT_PAL_STATUS_UNIMPLEMENTED(result);
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_platform_addr(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
struct ia64_pal_retval result;
|
|
|
|
INIT_PAL_STATUS_SUCCESS(result);
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_proc_get_features(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
struct ia64_pal_retval result = {0, 0, 0, 0};
|
|
long in0, in1, in2, in3;
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
result.status = ia64_pal_proc_get_features(&result.v0, &result.v1,
|
|
&result.v2, in2);
|
|
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_register_info(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
struct ia64_pal_retval result = {0, 0, 0, 0};
|
|
long in0, in1, in2, in3;
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
result.status = ia64_pal_register_info(in1, &result.v1, &result.v2);
|
|
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_cache_info(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
pal_cache_config_info_t ci;
|
|
long status;
|
|
unsigned long in0, in1, in2, in3, r9, r10;
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
status = ia64_pal_cache_config_info(in1, in2, &ci);
|
|
r9 = ci.pcci_info_1.pcci1_data;
|
|
r10 = ci.pcci_info_2.pcci2_data;
|
|
return ((struct ia64_pal_retval){status, r9, r10, 0});
|
|
}
|
|
|
|
#define GUEST_IMPL_VA_MSB 59
|
|
#define GUEST_RID_BITS 18
|
|
|
|
static struct ia64_pal_retval pal_vm_summary(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
pal_vm_info_1_u_t vminfo1;
|
|
pal_vm_info_2_u_t vminfo2;
|
|
struct ia64_pal_retval result;
|
|
|
|
PAL_CALL(result, PAL_VM_SUMMARY, 0, 0, 0);
|
|
if (!result.status) {
|
|
vminfo1.pvi1_val = result.v0;
|
|
vminfo1.pal_vm_info_1_s.max_itr_entry = 8;
|
|
vminfo1.pal_vm_info_1_s.max_dtr_entry = 8;
|
|
result.v0 = vminfo1.pvi1_val;
|
|
vminfo2.pal_vm_info_2_s.impl_va_msb = GUEST_IMPL_VA_MSB;
|
|
vminfo2.pal_vm_info_2_s.rid_size = GUEST_RID_BITS;
|
|
result.v1 = vminfo2.pvi2_val;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_vm_info(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct ia64_pal_retval result;
|
|
unsigned long in0, in1, in2, in3;
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
|
|
result.status = ia64_pal_vm_info(in1, in2,
|
|
(pal_tc_info_u_t *)&result.v1, &result.v2);
|
|
|
|
return result;
|
|
}
|
|
|
|
static u64 kvm_get_pal_call_index(struct kvm_vcpu *vcpu)
|
|
{
|
|
u64 index = 0;
|
|
struct exit_ctl_data *p;
|
|
|
|
p = kvm_get_exit_data(vcpu);
|
|
if (p->exit_reason == EXIT_REASON_PAL_CALL)
|
|
index = p->u.pal_data.gr28;
|
|
|
|
return index;
|
|
}
|
|
|
|
static void prepare_for_halt(struct kvm_vcpu *vcpu)
|
|
{
|
|
vcpu->arch.timer_pending = 1;
|
|
vcpu->arch.timer_fired = 0;
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_perf_mon_info(struct kvm_vcpu *vcpu)
|
|
{
|
|
long status;
|
|
unsigned long in0, in1, in2, in3, r9;
|
|
unsigned long pm_buffer[16];
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
status = ia64_pal_perf_mon_info(pm_buffer,
|
|
(pal_perf_mon_info_u_t *) &r9);
|
|
if (status != 0) {
|
|
printk(KERN_DEBUG"PAL_PERF_MON_INFO fails ret=%ld\n", status);
|
|
} else {
|
|
if (in1)
|
|
memcpy((void *)in1, pm_buffer, sizeof(pm_buffer));
|
|
else {
|
|
status = PAL_STATUS_EINVAL;
|
|
printk(KERN_WARNING"Invalid parameters "
|
|
"for PAL call:0x%lx!\n", in0);
|
|
}
|
|
}
|
|
return (struct ia64_pal_retval){status, r9, 0, 0};
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_halt_info(struct kvm_vcpu *vcpu)
|
|
{
|
|
unsigned long in0, in1, in2, in3;
|
|
long status;
|
|
unsigned long res = 1000UL | (1000UL << 16) | (10UL << 32)
|
|
| (1UL << 61) | (1UL << 60);
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
if (in1) {
|
|
memcpy((void *)in1, &res, sizeof(res));
|
|
status = 0;
|
|
} else{
|
|
status = PAL_STATUS_EINVAL;
|
|
printk(KERN_WARNING"Invalid parameters "
|
|
"for PAL call:0x%lx!\n", in0);
|
|
}
|
|
|
|
return (struct ia64_pal_retval){status, 0, 0, 0};
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_mem_attrib(struct kvm_vcpu *vcpu)
|
|
{
|
|
unsigned long r9;
|
|
long status;
|
|
|
|
status = ia64_pal_mem_attrib(&r9);
|
|
|
|
return (struct ia64_pal_retval){status, r9, 0, 0};
|
|
}
|
|
|
|
static void remote_pal_prefetch_visibility(void *v)
|
|
{
|
|
s64 trans_type = (s64)v;
|
|
ia64_pal_prefetch_visibility(trans_type);
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_prefetch_visibility(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct ia64_pal_retval result = {0, 0, 0, 0};
|
|
unsigned long in0, in1, in2, in3;
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
result.status = ia64_pal_prefetch_visibility(in1);
|
|
if (result.status == 0) {
|
|
/* Must be performed on all remote processors
|
|
in the coherence domain. */
|
|
smp_call_function(remote_pal_prefetch_visibility,
|
|
(void *)in1, 1);
|
|
/* Unnecessary on remote processor for other vcpus!*/
|
|
result.status = 1;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
static void remote_pal_mc_drain(void *v)
|
|
{
|
|
ia64_pal_mc_drain();
|
|
}
|
|
|
|
static struct ia64_pal_retval pal_get_brand_info(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct ia64_pal_retval result = {0, 0, 0, 0};
|
|
unsigned long in0, in1, in2, in3;
|
|
|
|
kvm_get_pal_call_data(vcpu, &in0, &in1, &in2, &in3);
|
|
|
|
if (in1 == 0 && in2) {
|
|
char brand_info[128];
|
|
result.status = ia64_pal_get_brand_info(brand_info);
|
|
if (result.status == PAL_STATUS_SUCCESS)
|
|
memcpy((void *)in2, brand_info, 128);
|
|
} else {
|
|
result.status = PAL_STATUS_REQUIRES_MEMORY;
|
|
printk(KERN_WARNING"Invalid parameters for "
|
|
"PAL call:0x%lx!\n", in0);
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
int kvm_pal_emul(struct kvm_vcpu *vcpu, struct kvm_run *run)
|
|
{
|
|
|
|
u64 gr28;
|
|
struct ia64_pal_retval result;
|
|
int ret = 1;
|
|
|
|
gr28 = kvm_get_pal_call_index(vcpu);
|
|
switch (gr28) {
|
|
case PAL_CACHE_FLUSH:
|
|
result = pal_cache_flush(vcpu);
|
|
break;
|
|
case PAL_MEM_ATTRIB:
|
|
result = pal_mem_attrib(vcpu);
|
|
break;
|
|
case PAL_CACHE_SUMMARY:
|
|
result = pal_cache_summary(vcpu);
|
|
break;
|
|
case PAL_PERF_MON_INFO:
|
|
result = pal_perf_mon_info(vcpu);
|
|
break;
|
|
case PAL_HALT_INFO:
|
|
result = pal_halt_info(vcpu);
|
|
break;
|
|
case PAL_HALT_LIGHT:
|
|
{
|
|
INIT_PAL_STATUS_SUCCESS(result);
|
|
prepare_for_halt(vcpu);
|
|
if (kvm_highest_pending_irq(vcpu) == -1)
|
|
ret = kvm_emulate_halt(vcpu);
|
|
}
|
|
break;
|
|
|
|
case PAL_PREFETCH_VISIBILITY:
|
|
result = pal_prefetch_visibility(vcpu);
|
|
break;
|
|
case PAL_MC_DRAIN:
|
|
result.status = ia64_pal_mc_drain();
|
|
/* FIXME: All vcpus likely call PAL_MC_DRAIN.
|
|
That causes the congestion. */
|
|
smp_call_function(remote_pal_mc_drain, NULL, 1);
|
|
break;
|
|
|
|
case PAL_FREQ_RATIOS:
|
|
result = pal_freq_ratios(vcpu);
|
|
break;
|
|
|
|
case PAL_FREQ_BASE:
|
|
result = pal_freq_base(vcpu);
|
|
break;
|
|
|
|
case PAL_LOGICAL_TO_PHYSICAL :
|
|
result = pal_logical_to_physica(vcpu);
|
|
break;
|
|
|
|
case PAL_VM_SUMMARY :
|
|
result = pal_vm_summary(vcpu);
|
|
break;
|
|
|
|
case PAL_VM_INFO :
|
|
result = pal_vm_info(vcpu);
|
|
break;
|
|
case PAL_PLATFORM_ADDR :
|
|
result = pal_platform_addr(vcpu);
|
|
break;
|
|
case PAL_CACHE_INFO:
|
|
result = pal_cache_info(vcpu);
|
|
break;
|
|
case PAL_PTCE_INFO:
|
|
INIT_PAL_STATUS_SUCCESS(result);
|
|
result.v1 = (1L << 32) | 1L;
|
|
break;
|
|
case PAL_REGISTER_INFO:
|
|
result = pal_register_info(vcpu);
|
|
break;
|
|
case PAL_VM_PAGE_SIZE:
|
|
result.status = ia64_pal_vm_page_size(&result.v0,
|
|
&result.v1);
|
|
break;
|
|
case PAL_RSE_INFO:
|
|
result.status = ia64_pal_rse_info(&result.v0,
|
|
(pal_hints_u_t *)&result.v1);
|
|
break;
|
|
case PAL_PROC_GET_FEATURES:
|
|
result = pal_proc_get_features(vcpu);
|
|
break;
|
|
case PAL_DEBUG_INFO:
|
|
result.status = ia64_pal_debug_info(&result.v0,
|
|
&result.v1);
|
|
break;
|
|
case PAL_VERSION:
|
|
result.status = ia64_pal_version(
|
|
(pal_version_u_t *)&result.v0,
|
|
(pal_version_u_t *)&result.v1);
|
|
break;
|
|
case PAL_FIXED_ADDR:
|
|
result.status = PAL_STATUS_SUCCESS;
|
|
result.v0 = vcpu->vcpu_id;
|
|
break;
|
|
case PAL_BRAND_INFO:
|
|
result = pal_get_brand_info(vcpu);
|
|
break;
|
|
case PAL_GET_PSTATE:
|
|
case PAL_CACHE_SHARED_INFO:
|
|
INIT_PAL_STATUS_UNIMPLEMENTED(result);
|
|
break;
|
|
default:
|
|
INIT_PAL_STATUS_UNIMPLEMENTED(result);
|
|
printk(KERN_WARNING"kvm: Unsupported pal call,"
|
|
" index:0x%lx\n", gr28);
|
|
}
|
|
set_pal_result(vcpu, result);
|
|
return ret;
|
|
}
|
|
|
|
static struct sal_ret_values sal_emulator(struct kvm *kvm,
|
|
long index, unsigned long in1,
|
|
unsigned long in2, unsigned long in3,
|
|
unsigned long in4, unsigned long in5,
|
|
unsigned long in6, unsigned long in7)
|
|
{
|
|
unsigned long r9 = 0;
|
|
unsigned long r10 = 0;
|
|
long r11 = 0;
|
|
long status;
|
|
|
|
status = 0;
|
|
switch (index) {
|
|
case SAL_FREQ_BASE:
|
|
status = ia64_sal_freq_base(in1, &r9, &r10);
|
|
break;
|
|
case SAL_PCI_CONFIG_READ:
|
|
printk(KERN_WARNING"kvm: Not allowed to call here!"
|
|
" SAL_PCI_CONFIG_READ\n");
|
|
break;
|
|
case SAL_PCI_CONFIG_WRITE:
|
|
printk(KERN_WARNING"kvm: Not allowed to call here!"
|
|
" SAL_PCI_CONFIG_WRITE\n");
|
|
break;
|
|
case SAL_SET_VECTORS:
|
|
if (in1 == SAL_VECTOR_OS_BOOT_RENDEZ) {
|
|
if (in4 != 0 || in5 != 0 || in6 != 0 || in7 != 0) {
|
|
status = -2;
|
|
} else {
|
|
kvm->arch.rdv_sal_data.boot_ip = in2;
|
|
kvm->arch.rdv_sal_data.boot_gp = in3;
|
|
}
|
|
printk("Rendvous called! iip:%lx\n\n", in2);
|
|
} else
|
|
printk(KERN_WARNING"kvm: CALLED SAL_SET_VECTORS %lu."
|
|
"ignored...\n", in1);
|
|
break;
|
|
case SAL_GET_STATE_INFO:
|
|
/* No more info. */
|
|
status = -5;
|
|
r9 = 0;
|
|
break;
|
|
case SAL_GET_STATE_INFO_SIZE:
|
|
/* Return a dummy size. */
|
|
status = 0;
|
|
r9 = 128;
|
|
break;
|
|
case SAL_CLEAR_STATE_INFO:
|
|
/* Noop. */
|
|
break;
|
|
case SAL_MC_RENDEZ:
|
|
printk(KERN_WARNING
|
|
"kvm: called SAL_MC_RENDEZ. ignored...\n");
|
|
break;
|
|
case SAL_MC_SET_PARAMS:
|
|
printk(KERN_WARNING
|
|
"kvm: called SAL_MC_SET_PARAMS.ignored!\n");
|
|
break;
|
|
case SAL_CACHE_FLUSH:
|
|
if (1) {
|
|
/*Flush using SAL.
|
|
This method is faster but has a side
|
|
effect on other vcpu running on
|
|
this cpu. */
|
|
status = ia64_sal_cache_flush(in1);
|
|
} else {
|
|
/*Maybe need to implement the method
|
|
without side effect!*/
|
|
status = 0;
|
|
}
|
|
break;
|
|
case SAL_CACHE_INIT:
|
|
printk(KERN_WARNING
|
|
"kvm: called SAL_CACHE_INIT. ignored...\n");
|
|
break;
|
|
case SAL_UPDATE_PAL:
|
|
printk(KERN_WARNING
|
|
"kvm: CALLED SAL_UPDATE_PAL. ignored...\n");
|
|
break;
|
|
default:
|
|
printk(KERN_WARNING"kvm: called SAL_CALL with unknown index."
|
|
" index:%ld\n", index);
|
|
status = -1;
|
|
break;
|
|
}
|
|
return ((struct sal_ret_values) {status, r9, r10, r11});
|
|
}
|
|
|
|
static void kvm_get_sal_call_data(struct kvm_vcpu *vcpu, u64 *in0, u64 *in1,
|
|
u64 *in2, u64 *in3, u64 *in4, u64 *in5, u64 *in6, u64 *in7){
|
|
|
|
struct exit_ctl_data *p;
|
|
|
|
p = kvm_get_exit_data(vcpu);
|
|
|
|
if (p->exit_reason == EXIT_REASON_SAL_CALL) {
|
|
*in0 = p->u.sal_data.in0;
|
|
*in1 = p->u.sal_data.in1;
|
|
*in2 = p->u.sal_data.in2;
|
|
*in3 = p->u.sal_data.in3;
|
|
*in4 = p->u.sal_data.in4;
|
|
*in5 = p->u.sal_data.in5;
|
|
*in6 = p->u.sal_data.in6;
|
|
*in7 = p->u.sal_data.in7;
|
|
return ;
|
|
}
|
|
*in0 = 0;
|
|
}
|
|
|
|
void kvm_sal_emul(struct kvm_vcpu *vcpu)
|
|
{
|
|
|
|
struct sal_ret_values result;
|
|
u64 index, in1, in2, in3, in4, in5, in6, in7;
|
|
|
|
kvm_get_sal_call_data(vcpu, &index, &in1, &in2,
|
|
&in3, &in4, &in5, &in6, &in7);
|
|
result = sal_emulator(vcpu->kvm, index, in1, in2, in3,
|
|
in4, in5, in6, in7);
|
|
set_sal_result(vcpu, result);
|
|
}
|