2 * Kernel-based Virtual Machine driver for Linux
4 * derived from drivers/kvm/kvm_main.c
6 * Copyright (C) 2006 Qumranet, Inc.
9 * Avi Kivity <avi@qumranet.com>
10 * Yaniv Kamay <yaniv@qumranet.com>
12 * This work is licensed under the terms of the GNU GPL, version 2. See
13 * the COPYING file in the top-level directory.
21 #include <linux/kvm.h>
23 #include <linux/vmalloc.h>
25 #include <asm/uaccess.h>
27 #define MAX_IO_MSRS 256
30 * List of msr numbers which we expose to userspace through KVM_GET_MSRS
31 * and KVM_SET_MSRS, and KVM_GET_MSR_INDEX_LIST.
33 * This list is modified at module load time to reflect the
34 * capabilities of the host cpu.
36 static u32 msrs_to_save[] = {
37 MSR_IA32_SYSENTER_CS, MSR_IA32_SYSENTER_ESP, MSR_IA32_SYSENTER_EIP,
40 MSR_CSTAR, MSR_KERNEL_GS_BASE, MSR_SYSCALL_MASK, MSR_LSTAR,
42 MSR_IA32_TIME_STAMP_COUNTER,
45 static unsigned num_msrs_to_save;
47 static u32 emulated_msrs[] = {
52 * Adapt set_msr() to msr_io()'s calling convention
54 static int do_set_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data)
56 return kvm_set_msr(vcpu, index, *data);
60 * Read or write a bunch of msrs. All parameters are kernel addresses.
62 * @return number of msrs set successfully.
64 static int __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs,
65 struct kvm_msr_entry *entries,
66 int (*do_msr)(struct kvm_vcpu *vcpu,
67 unsigned index, u64 *data))
73 for (i = 0; i < msrs->nmsrs; ++i)
74 if (do_msr(vcpu, entries[i].index, &entries[i].data))
83 * Read or write a bunch of msrs. Parameters are user addresses.
85 * @return number of msrs set successfully.
87 static int msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs __user *user_msrs,
88 int (*do_msr)(struct kvm_vcpu *vcpu,
89 unsigned index, u64 *data),
93 struct kvm_msr_entry *entries;
98 if (copy_from_user(&msrs, user_msrs, sizeof msrs))
102 if (msrs.nmsrs >= MAX_IO_MSRS)
106 size = sizeof(struct kvm_msr_entry) * msrs.nmsrs;
107 entries = vmalloc(size);
112 if (copy_from_user(entries, user_msrs->entries, size))
115 r = n = __msr_io(vcpu, &msrs, entries, do_msr);
120 if (writeback && copy_to_user(user_msrs->entries, entries, size))
131 long kvm_arch_dev_ioctl(struct file *filp,
132 unsigned int ioctl, unsigned long arg)
134 void __user *argp = (void __user *)arg;
138 case KVM_GET_MSR_INDEX_LIST: {
139 struct kvm_msr_list __user *user_msr_list = argp;
140 struct kvm_msr_list msr_list;
144 if (copy_from_user(&msr_list, user_msr_list, sizeof msr_list))
147 msr_list.nmsrs = num_msrs_to_save + ARRAY_SIZE(emulated_msrs);
148 if (copy_to_user(user_msr_list, &msr_list, sizeof msr_list))
151 if (n < num_msrs_to_save)
154 if (copy_to_user(user_msr_list->indices, &msrs_to_save,
155 num_msrs_to_save * sizeof(u32)))
157 if (copy_to_user(user_msr_list->indices
158 + num_msrs_to_save * sizeof(u32),
160 ARRAY_SIZE(emulated_msrs) * sizeof(u32)))
172 void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
174 kvm_x86_ops->vcpu_load(vcpu, cpu);
177 void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
179 kvm_x86_ops->vcpu_put(vcpu);
182 static void cpuid_fix_nx_cap(struct kvm_vcpu *vcpu)
186 struct kvm_cpuid_entry *e, *entry;
188 rdmsrl(MSR_EFER, efer);
190 for (i = 0; i < vcpu->cpuid_nent; ++i) {
191 e = &vcpu->cpuid_entries[i];
192 if (e->function == 0x80000001) {
197 if (entry && (entry->edx & (1 << 20)) && !(efer & EFER_NX)) {
198 entry->edx &= ~(1 << 20);
199 printk(KERN_INFO "kvm: guest NX capability removed\n");
203 static int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
204 struct kvm_cpuid *cpuid,
205 struct kvm_cpuid_entry __user *entries)
210 if (cpuid->nent > KVM_MAX_CPUID_ENTRIES)
213 if (copy_from_user(&vcpu->cpuid_entries, entries,
214 cpuid->nent * sizeof(struct kvm_cpuid_entry)))
216 vcpu->cpuid_nent = cpuid->nent;
217 cpuid_fix_nx_cap(vcpu);
224 static int kvm_vcpu_ioctl_get_lapic(struct kvm_vcpu *vcpu,
225 struct kvm_lapic_state *s)
228 memcpy(s->regs, vcpu->apic->regs, sizeof *s);
234 static int kvm_vcpu_ioctl_set_lapic(struct kvm_vcpu *vcpu,
235 struct kvm_lapic_state *s)
238 memcpy(vcpu->apic->regs, s->regs, sizeof *s);
239 kvm_apic_post_state_restore(vcpu);
245 long kvm_arch_vcpu_ioctl(struct file *filp,
246 unsigned int ioctl, unsigned long arg)
248 struct kvm_vcpu *vcpu = filp->private_data;
249 void __user *argp = (void __user *)arg;
253 case KVM_GET_LAPIC: {
254 struct kvm_lapic_state lapic;
256 memset(&lapic, 0, sizeof lapic);
257 r = kvm_vcpu_ioctl_get_lapic(vcpu, &lapic);
261 if (copy_to_user(argp, &lapic, sizeof lapic))
266 case KVM_SET_LAPIC: {
267 struct kvm_lapic_state lapic;
270 if (copy_from_user(&lapic, argp, sizeof lapic))
272 r = kvm_vcpu_ioctl_set_lapic(vcpu, &lapic);;
278 case KVM_SET_CPUID: {
279 struct kvm_cpuid __user *cpuid_arg = argp;
280 struct kvm_cpuid cpuid;
283 if (copy_from_user(&cpuid, cpuid_arg, sizeof cpuid))
285 r = kvm_vcpu_ioctl_set_cpuid(vcpu, &cpuid, cpuid_arg->entries);
291 r = msr_io(vcpu, argp, kvm_get_msr, 1);
294 r = msr_io(vcpu, argp, do_set_msr, 0);
303 static int kvm_vm_ioctl_set_tss_addr(struct kvm *kvm, unsigned long addr)
307 if (addr > (unsigned int)(-3 * PAGE_SIZE))
309 ret = kvm_x86_ops->set_tss_addr(kvm, addr);
313 static int kvm_vm_ioctl_set_nr_mmu_pages(struct kvm *kvm,
314 u32 kvm_nr_mmu_pages)
316 if (kvm_nr_mmu_pages < KVM_MIN_ALLOC_MMU_PAGES)
319 mutex_lock(&kvm->lock);
321 kvm_mmu_change_mmu_pages(kvm, kvm_nr_mmu_pages);
322 kvm->n_requested_mmu_pages = kvm_nr_mmu_pages;
324 mutex_unlock(&kvm->lock);
328 static int kvm_vm_ioctl_get_nr_mmu_pages(struct kvm *kvm)
330 return kvm->n_alloc_mmu_pages;
334 * Set a new alias region. Aliases map a portion of physical memory into
335 * another portion. This is useful for memory windows, for example the PC
338 static int kvm_vm_ioctl_set_memory_alias(struct kvm *kvm,
339 struct kvm_memory_alias *alias)
342 struct kvm_mem_alias *p;
345 /* General sanity checks */
346 if (alias->memory_size & (PAGE_SIZE - 1))
348 if (alias->guest_phys_addr & (PAGE_SIZE - 1))
350 if (alias->slot >= KVM_ALIAS_SLOTS)
352 if (alias->guest_phys_addr + alias->memory_size
353 < alias->guest_phys_addr)
355 if (alias->target_phys_addr + alias->memory_size
356 < alias->target_phys_addr)
359 mutex_lock(&kvm->lock);
361 p = &kvm->aliases[alias->slot];
362 p->base_gfn = alias->guest_phys_addr >> PAGE_SHIFT;
363 p->npages = alias->memory_size >> PAGE_SHIFT;
364 p->target_gfn = alias->target_phys_addr >> PAGE_SHIFT;
366 for (n = KVM_ALIAS_SLOTS; n > 0; --n)
367 if (kvm->aliases[n - 1].npages)
371 kvm_mmu_zap_all(kvm);
373 mutex_unlock(&kvm->lock);
381 static int kvm_vm_ioctl_get_irqchip(struct kvm *kvm, struct kvm_irqchip *chip)
386 switch (chip->chip_id) {
387 case KVM_IRQCHIP_PIC_MASTER:
388 memcpy(&chip->chip.pic,
389 &pic_irqchip(kvm)->pics[0],
390 sizeof(struct kvm_pic_state));
392 case KVM_IRQCHIP_PIC_SLAVE:
393 memcpy(&chip->chip.pic,
394 &pic_irqchip(kvm)->pics[1],
395 sizeof(struct kvm_pic_state));
397 case KVM_IRQCHIP_IOAPIC:
398 memcpy(&chip->chip.ioapic,
400 sizeof(struct kvm_ioapic_state));
409 static int kvm_vm_ioctl_set_irqchip(struct kvm *kvm, struct kvm_irqchip *chip)
414 switch (chip->chip_id) {
415 case KVM_IRQCHIP_PIC_MASTER:
416 memcpy(&pic_irqchip(kvm)->pics[0],
418 sizeof(struct kvm_pic_state));
420 case KVM_IRQCHIP_PIC_SLAVE:
421 memcpy(&pic_irqchip(kvm)->pics[1],
423 sizeof(struct kvm_pic_state));
425 case KVM_IRQCHIP_IOAPIC:
426 memcpy(ioapic_irqchip(kvm),
428 sizeof(struct kvm_ioapic_state));
434 kvm_pic_update_irq(pic_irqchip(kvm));
438 long kvm_arch_vm_ioctl(struct file *filp,
439 unsigned int ioctl, unsigned long arg)
441 struct kvm *kvm = filp->private_data;
442 void __user *argp = (void __user *)arg;
446 case KVM_SET_TSS_ADDR:
447 r = kvm_vm_ioctl_set_tss_addr(kvm, arg);
451 case KVM_SET_MEMORY_REGION: {
452 struct kvm_memory_region kvm_mem;
453 struct kvm_userspace_memory_region kvm_userspace_mem;
456 if (copy_from_user(&kvm_mem, argp, sizeof kvm_mem))
458 kvm_userspace_mem.slot = kvm_mem.slot;
459 kvm_userspace_mem.flags = kvm_mem.flags;
460 kvm_userspace_mem.guest_phys_addr = kvm_mem.guest_phys_addr;
461 kvm_userspace_mem.memory_size = kvm_mem.memory_size;
462 r = kvm_vm_ioctl_set_memory_region(kvm, &kvm_userspace_mem, 0);
467 case KVM_SET_NR_MMU_PAGES:
468 r = kvm_vm_ioctl_set_nr_mmu_pages(kvm, arg);
472 case KVM_GET_NR_MMU_PAGES:
473 r = kvm_vm_ioctl_get_nr_mmu_pages(kvm);
475 case KVM_SET_MEMORY_ALIAS: {
476 struct kvm_memory_alias alias;
479 if (copy_from_user(&alias, argp, sizeof alias))
481 r = kvm_vm_ioctl_set_memory_alias(kvm, &alias);
486 case KVM_CREATE_IRQCHIP:
488 kvm->vpic = kvm_create_pic(kvm);
490 r = kvm_ioapic_init(kvm);
500 struct kvm_irq_level irq_event;
503 if (copy_from_user(&irq_event, argp, sizeof irq_event))
505 if (irqchip_in_kernel(kvm)) {
506 mutex_lock(&kvm->lock);
507 if (irq_event.irq < 16)
508 kvm_pic_set_irq(pic_irqchip(kvm),
511 kvm_ioapic_set_irq(kvm->vioapic,
514 mutex_unlock(&kvm->lock);
519 case KVM_GET_IRQCHIP: {
520 /* 0: PIC master, 1: PIC slave, 2: IOAPIC */
521 struct kvm_irqchip chip;
524 if (copy_from_user(&chip, argp, sizeof chip))
527 if (!irqchip_in_kernel(kvm))
529 r = kvm_vm_ioctl_get_irqchip(kvm, &chip);
533 if (copy_to_user(argp, &chip, sizeof chip))
538 case KVM_SET_IRQCHIP: {
539 /* 0: PIC master, 1: PIC slave, 2: IOAPIC */
540 struct kvm_irqchip chip;
543 if (copy_from_user(&chip, argp, sizeof chip))
546 if (!irqchip_in_kernel(kvm))
548 r = kvm_vm_ioctl_set_irqchip(kvm, &chip);
561 static __init void kvm_init_msr_list(void)
566 for (i = j = 0; i < ARRAY_SIZE(msrs_to_save); i++) {
567 if (rdmsr_safe(msrs_to_save[i], &dummy[0], &dummy[1]) < 0)
570 msrs_to_save[j] = msrs_to_save[i];
573 num_msrs_to_save = j;
576 __init void kvm_arch_init(void)