| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * tools/testing/selftests/kvm/include/memstress.h |
| * |
| * Copyright (C) 2020, Google LLC. |
| */ |
| |
| #ifndef SELFTEST_KVM_MEMSTRESS_H |
| #define SELFTEST_KVM_MEMSTRESS_H |
| |
| #include <pthread.h> |
| |
| #include "kvm_util.h" |
| |
| /* Default guest test virtual memory offset */ |
| #define DEFAULT_GUEST_TEST_MEM 0xc0000000 |
| |
| #define DEFAULT_PER_VCPU_MEM_SIZE (1 << 30) /* 1G */ |
| |
| #define MEMSTRESS_MEM_SLOT_INDEX 1 |
| |
| struct memstress_vcpu_args { |
| uint64_t gpa; |
| uint64_t gva; |
| uint64_t pages; |
| |
| /* Only used by the host userspace part of the vCPU thread */ |
| struct kvm_vcpu *vcpu; |
| int vcpu_idx; |
| }; |
| |
| struct memstress_args { |
| struct kvm_vm *vm; |
| /* The starting address and size of the guest test region. */ |
| uint64_t gpa; |
| uint64_t size; |
| uint64_t guest_page_size; |
| uint32_t random_seed; |
| uint32_t write_percent; |
| |
| /* Run vCPUs in L2 instead of L1, if the architecture supports it. */ |
| bool nested; |
| /* Randomize which pages are accessed by the guest. */ |
| bool random_access; |
| /* True if all vCPUs are pinned to pCPUs */ |
| bool pin_vcpus; |
| /* The vCPU=>pCPU pinning map. Only valid if pin_vcpus is true. */ |
| uint32_t vcpu_to_pcpu[KVM_MAX_VCPUS]; |
| |
| /* Test is done, stop running vCPUs. */ |
| bool stop_vcpus; |
| |
| struct memstress_vcpu_args vcpu_args[KVM_MAX_VCPUS]; |
| }; |
| |
| extern struct memstress_args memstress_args; |
| |
| struct kvm_vm *memstress_create_vm(enum vm_guest_mode mode, int nr_vcpus, |
| uint64_t vcpu_memory_bytes, int slots, |
| enum vm_mem_backing_src_type backing_src, |
| bool partition_vcpu_memory_access); |
| void memstress_destroy_vm(struct kvm_vm *vm); |
| |
| void memstress_set_write_percent(struct kvm_vm *vm, uint32_t write_percent); |
| void memstress_set_random_seed(struct kvm_vm *vm, uint32_t random_seed); |
| void memstress_set_random_access(struct kvm_vm *vm, bool random_access); |
| |
| void memstress_start_vcpu_threads(int vcpus, void (*vcpu_fn)(struct memstress_vcpu_args *)); |
| void memstress_join_vcpu_threads(int vcpus); |
| void memstress_guest_code(uint32_t vcpu_id); |
| |
| uint64_t memstress_nested_pages(int nr_vcpus); |
| void memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu *vcpus[]); |
| |
| void memstress_enable_dirty_logging(struct kvm_vm *vm, int slots); |
| void memstress_disable_dirty_logging(struct kvm_vm *vm, int slots); |
| void memstress_get_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], int slots); |
| void memstress_clear_dirty_log(struct kvm_vm *vm, unsigned long *bitmaps[], |
| int slots, uint64_t pages_per_slot); |
| unsigned long **memstress_alloc_bitmaps(int slots, uint64_t pages_per_slot); |
| void memstress_free_bitmaps(unsigned long *bitmaps[], int slots); |
| |
| #endif /* SELFTEST_KVM_MEMSTRESS_H */ |