On Mon, Sep 4, 2023 at 9:24 PM Andrew Jones <ajones@xxxxxxxxxxxxxxxx> wrote: > > On Sat, Sep 02, 2023 at 08:59:24PM +0800, Haibo Xu wrote: > > Split the arch-neutral test code out of aarch64/arch_timer.c > > and put them into a common arch_timer.c. This is a preparation > > to share timer test codes in riscv. > > > > Suggested-by: Andrew Jones <ajones@xxxxxxxxxxxxxxxx> > > Signed-off-by: Haibo Xu <haibo1.xu@xxxxxxxxx> > > --- > > tools/testing/selftests/kvm/Makefile | 9 +- > > .../selftests/kvm/aarch64/arch_timer.c | 288 +----------------- > > tools/testing/selftests/kvm/arch_timer.c | 252 +++++++++++++++ > > .../selftests/kvm/include/timer_test.h | 52 ++++ > > 4 files changed, 317 insertions(+), 284 deletions(-) > > create mode 100644 tools/testing/selftests/kvm/arch_timer.c > > create mode 100644 tools/testing/selftests/kvm/include/timer_test.h > > > > diff --git a/tools/testing/selftests/kvm/Makefile b/tools/testing/selftests/kvm/Makefile > > index 0b9c42fbce8c..fb8904e2c06a 100644 > > --- a/tools/testing/selftests/kvm/Makefile > > +++ b/tools/testing/selftests/kvm/Makefile > > @@ -140,7 +140,6 @@ TEST_GEN_PROGS_x86_64 += system_counter_offset_test > > TEST_GEN_PROGS_EXTENDED_x86_64 += x86_64/nx_huge_pages_test > > > > TEST_GEN_PROGS_aarch64 += aarch64/aarch32_id_regs > > -TEST_GEN_PROGS_aarch64 += aarch64/arch_timer > > TEST_GEN_PROGS_aarch64 += aarch64/debug-exceptions > > TEST_GEN_PROGS_aarch64 += aarch64/hypercalls > > TEST_GEN_PROGS_aarch64 += aarch64/page_fault_test > > @@ -150,6 +149,7 @@ TEST_GEN_PROGS_aarch64 += aarch64/vcpu_width_config > > TEST_GEN_PROGS_aarch64 += aarch64/vgic_init > > TEST_GEN_PROGS_aarch64 += aarch64/vgic_irq > > TEST_GEN_PROGS_aarch64 += access_tracking_perf_test > > +TEST_GEN_PROGS_aarch64 += arch_timer > > TEST_GEN_PROGS_aarch64 += demand_paging_test > > TEST_GEN_PROGS_aarch64 += dirty_log_test > > TEST_GEN_PROGS_aarch64 += dirty_log_perf_test > > @@ -188,6 +188,7 @@ TEST_GEN_PROGS_riscv += set_memory_region_test > > TEST_GEN_PROGS_riscv += kvm_binary_stats_test > > > > SPLIT_TESTS += get-reg-list > > +SPLIT_TESTS += arch_timer > > > > TEST_PROGS += $(TEST_PROGS_$(ARCH_DIR)) > > TEST_GEN_PROGS += $(TEST_GEN_PROGS_$(ARCH_DIR)) > > @@ -248,13 +249,10 @@ TEST_DEP_FILES += $(patsubst %.o, %.d, $(SPLIT_TESTS_OBJS)) > > -include $(TEST_DEP_FILES) > > > > $(TEST_GEN_PROGS) $(TEST_GEN_PROGS_EXTENDED): %: %.o > > - $(CC) $(CFLAGS) $(CPPFLAGS) $(LDFLAGS) $(TARGET_ARCH) $< $(LIBKVM_OBJS) $(LDLIBS) -o $@ > > + $(CC) $(CFLAGS) $(CPPFLAGS) $(LDFLAGS) $(TARGET_ARCH) $^ $(LDLIBS) -o $@ > > $(TEST_GEN_OBJ): $(OUTPUT)/%.o: %.c > > $(CC) $(CFLAGS) $(CPPFLAGS) $(TARGET_ARCH) -c $< -o $@ > > > > -$(SPLIT_TESTS_TARGETS): %: %.o $(SPLIT_TESTS_OBJS) > > - $(CC) $(CFLAGS) $(CPPFLAGS) $(LDFLAGS) $(TARGET_ARCH) $^ $(LDLIBS) -o $@ > > - > > EXTRA_CLEAN += $(LIBKVM_OBJS) $(TEST_DEP_FILES) $(TEST_GEN_OBJ) $(SPLIT_TESTS_OBJS) cscope.* > > > > x := $(shell mkdir -p $(sort $(dir $(LIBKVM_C_OBJ) $(LIBKVM_S_OBJ)))) > > @@ -273,6 +271,7 @@ $(LIBKVM_STRING_OBJ): $(OUTPUT)/%.o: %.c > > x := $(shell mkdir -p $(sort $(dir $(TEST_GEN_PROGS)))) > > $(TEST_GEN_PROGS): $(LIBKVM_OBJS) > > $(TEST_GEN_PROGS_EXTENDED): $(LIBKVM_OBJS) > > +$(SPLIT_TESTS_TARGETS): $(OUTPUT)/%: $(ARCH_DIR)/%.o > > The improvements to the Makefile to avoid SPLIT_TESTS_TARGETS needing its > own $(CC) line should preferably be done in a separate, preliminary patch. > Yes, Will move the change to a separate patch. > > > > cscope: include_paths = $(LINUX_TOOL_INCLUDE) $(LINUX_HDR_PATH) include lib .. > > cscope: > > diff --git a/tools/testing/selftests/kvm/aarch64/arch_timer.c b/tools/testing/selftests/kvm/aarch64/arch_timer.c > > index b63859829a96..ceb649548751 100644 > > --- a/tools/testing/selftests/kvm/aarch64/arch_timer.c > > +++ b/tools/testing/selftests/kvm/aarch64/arch_timer.c > > @@ -1,91 +1,25 @@ > > // SPDX-License-Identifier: GPL-2.0-only > > /* > > - * arch_timer.c - Tests the aarch64 timer IRQ functionality > > - * > > * The test validates both the virtual and physical timer IRQs using > > - * CVAL and TVAL registers. This consitutes the four stages in the test. > > - * The guest's main thread configures the timer interrupt for a stage > > - * and waits for it to fire, with a timeout equal to the timer period. > > - * It asserts that the timeout doesn't exceed the timer period. > > - * > > - * On the other hand, upon receipt of an interrupt, the guest's interrupt > > - * handler validates the interrupt by checking if the architectural state > > - * is in compliance with the specifications. > > - * > > - * The test provides command-line options to configure the timer's > > - * period (-p), number of vCPUs (-n), and iterations per stage (-i). > > - * To stress-test the timer stack even more, an option to migrate the > > - * vCPUs across pCPUs (-m), at a particular rate, is also provided. > > + * CVAL and TVAL registers. > > * > > * Copyright (c) 2021, Google LLC. > > */ > > #define _GNU_SOURCE > > > > -#include <stdlib.h> > > -#include <pthread.h> > > -#include <linux/kvm.h> > > -#include <linux/sizes.h> > > -#include <linux/bitmap.h> > > -#include <sys/sysinfo.h> > > - > > -#include "kvm_util.h" > > -#include "processor.h" > > -#include "delay.h" > > #include "arch_timer.h" > > +#include "delay.h" > > #include "gic.h" > > +#include "processor.h" > > +#include "timer_test.h" > > #include "vgic.h" > > > > -#define NR_VCPUS_DEF 4 > > -#define NR_TEST_ITERS_DEF 5 > > -#define TIMER_TEST_PERIOD_MS_DEF 10 > > -#define TIMER_TEST_ERR_MARGIN_US 100 > > -#define TIMER_TEST_MIGRATION_FREQ_MS 2 > > - > > -struct test_args { > > - int nr_vcpus; > > - int nr_iter; > > - int timer_period_ms; > > - int migration_freq_ms; > > - struct kvm_arm_counter_offset offset; > > -}; > > - > > -static struct test_args test_args = { > > - .nr_vcpus = NR_VCPUS_DEF, > > - .nr_iter = NR_TEST_ITERS_DEF, > > - .timer_period_ms = TIMER_TEST_PERIOD_MS_DEF, > > - .migration_freq_ms = TIMER_TEST_MIGRATION_FREQ_MS, > > - .offset = { .reserved = 1 }, > > -}; > > - > > -#define msecs_to_usecs(msec) ((msec) * 1000LL) > > - > > -#define GICD_BASE_GPA 0x8000000ULL > > -#define GICR_BASE_GPA 0x80A0000ULL > > - > > -enum guest_stage { > > - GUEST_STAGE_VTIMER_CVAL = 1, > > - GUEST_STAGE_VTIMER_TVAL, > > - GUEST_STAGE_PTIMER_CVAL, > > - GUEST_STAGE_PTIMER_TVAL, > > - GUEST_STAGE_MAX, > > -}; > > - > > -/* Shared variables between host and guest */ > > -struct test_vcpu_shared_data { > > - int nr_iter; > > - enum guest_stage guest_stage; > > - uint64_t xcnt; > > -}; > > - > > -static struct kvm_vcpu *vcpus[KVM_MAX_VCPUS]; > > -static pthread_t pt_vcpu_run[KVM_MAX_VCPUS]; > > -static struct test_vcpu_shared_data vcpu_shared_data[KVM_MAX_VCPUS]; > > +extern struct test_args test_args; > > +extern struct kvm_vcpu *vcpus[]; > > +extern struct test_vcpu_shared_data vcpu_shared_data[]; > > > > static int vtimer_irq, ptimer_irq; > > > > -static unsigned long *vcpu_done_map; > > -static pthread_mutex_t vcpu_done_map_lock; > > - > > static void > > guest_configure_timer_action(struct test_vcpu_shared_data *shared_data) > > { > > @@ -222,137 +156,6 @@ static void guest_code(void) > > GUEST_DONE(); > > } > > > > -static void *test_vcpu_run(void *arg) > > -{ > > - unsigned int vcpu_idx = (unsigned long)arg; > > - struct ucall uc; > > - struct kvm_vcpu *vcpu = vcpus[vcpu_idx]; > > - struct kvm_vm *vm = vcpu->vm; > > - struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx]; > > - > > - vcpu_run(vcpu); > > - > > - /* Currently, any exit from guest is an indication of completion */ > > - pthread_mutex_lock(&vcpu_done_map_lock); > > - __set_bit(vcpu_idx, vcpu_done_map); > > - pthread_mutex_unlock(&vcpu_done_map_lock); > > - > > - switch (get_ucall(vcpu, &uc)) { > > - case UCALL_SYNC: > > - case UCALL_DONE: > > - break; > > - case UCALL_ABORT: > > - sync_global_from_guest(vm, *shared_data); > > - fprintf(stderr, "Guest assert failed, vcpu %u; stage; %u; iter: %u\n", > > - vcpu_idx, shared_data->guest_stage, shared_data->nr_iter); > > - REPORT_GUEST_ASSERT(uc); > > - break; > > - default: > > - TEST_FAIL("Unexpected guest exit\n"); > > - } > > - > > - return NULL; > > -} > > - > > -static uint32_t test_get_pcpu(void) > > -{ > > - uint32_t pcpu; > > - unsigned int nproc_conf; > > - cpu_set_t online_cpuset; > > - > > - nproc_conf = get_nprocs_conf(); > > - sched_getaffinity(0, sizeof(cpu_set_t), &online_cpuset); > > - > > - /* Randomly find an available pCPU to place a vCPU on */ > > - do { > > - pcpu = rand() % nproc_conf; > > - } while (!CPU_ISSET(pcpu, &online_cpuset)); > > - > > - return pcpu; > > -} > > - > > -static int test_migrate_vcpu(unsigned int vcpu_idx) > > -{ > > - int ret; > > - cpu_set_t cpuset; > > - uint32_t new_pcpu = test_get_pcpu(); > > - > > - CPU_ZERO(&cpuset); > > - CPU_SET(new_pcpu, &cpuset); > > - > > - pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu); > > - > > - ret = pthread_setaffinity_np(pt_vcpu_run[vcpu_idx], > > - sizeof(cpuset), &cpuset); > > - > > - /* Allow the error where the vCPU thread is already finished */ > > - TEST_ASSERT(ret == 0 || ret == ESRCH, > > - "Failed to migrate the vCPU:%u to pCPU: %u; ret: %d\n", > > - vcpu_idx, new_pcpu, ret); > > - > > - return ret; > > -} > > - > > -static void *test_vcpu_migration(void *arg) > > -{ > > - unsigned int i, n_done; > > - bool vcpu_done; > > - > > - do { > > - usleep(msecs_to_usecs(test_args.migration_freq_ms)); > > - > > - for (n_done = 0, i = 0; i < test_args.nr_vcpus; i++) { > > - pthread_mutex_lock(&vcpu_done_map_lock); > > - vcpu_done = test_bit(i, vcpu_done_map); > > - pthread_mutex_unlock(&vcpu_done_map_lock); > > - > > - if (vcpu_done) { > > - n_done++; > > - continue; > > - } > > - > > - test_migrate_vcpu(i); > > - } > > - } while (test_args.nr_vcpus != n_done); > > - > > - return NULL; > > -} > > - > > -static void test_run(struct kvm_vm *vm) > > -{ > > - pthread_t pt_vcpu_migration; > > - unsigned int i; > > - int ret; > > - > > - pthread_mutex_init(&vcpu_done_map_lock, NULL); > > - vcpu_done_map = bitmap_zalloc(test_args.nr_vcpus); > > - TEST_ASSERT(vcpu_done_map, "Failed to allocate vcpu done bitmap\n"); > > - > > - for (i = 0; i < (unsigned long)test_args.nr_vcpus; i++) { > > - ret = pthread_create(&pt_vcpu_run[i], NULL, test_vcpu_run, > > - (void *)(unsigned long)i); > > - TEST_ASSERT(!ret, "Failed to create vCPU-%d pthread\n", i); > > - } > > - > > - /* Spawn a thread to control the vCPU migrations */ > > - if (test_args.migration_freq_ms) { > > - srand(time(NULL)); > > - > > - ret = pthread_create(&pt_vcpu_migration, NULL, > > - test_vcpu_migration, NULL); > > - TEST_ASSERT(!ret, "Failed to create the migration pthread\n"); > > - } > > - > > - > > - for (i = 0; i < test_args.nr_vcpus; i++) > > - pthread_join(pt_vcpu_run[i], NULL); > > - > > - if (test_args.migration_freq_ms) > > - pthread_join(pt_vcpu_migration, NULL); > > - > > - bitmap_free(vcpu_done_map); > > -} > > - > > static void test_init_timer_irq(struct kvm_vm *vm) > > { > > /* Timer initid should be same for all the vCPUs, so query only vCPU-0 */ > > @@ -369,7 +172,7 @@ static void test_init_timer_irq(struct kvm_vm *vm) > > > > static int gic_fd; > > > > -static struct kvm_vm *test_vm_create(void) > > +struct kvm_vm *test_vm_create(void) > > { > > struct kvm_vm *vm; > > unsigned int i; > > @@ -400,81 +203,8 @@ static struct kvm_vm *test_vm_create(void) > > return vm; > > } > > > > -static void test_vm_cleanup(struct kvm_vm *vm) > > +void test_vm_cleanup(struct kvm_vm *vm) > > { > > close(gic_fd); > > kvm_vm_free(vm); > > } > > - > > -static void test_print_help(char *name) > > -{ > > - pr_info("Usage: %s [-h] [-n nr_vcpus] [-i iterations] [-p timer_period_ms]\n", > > - name); > > - pr_info("\t-n: Number of vCPUs to configure (default: %u; max: %u)\n", > > - NR_VCPUS_DEF, KVM_MAX_VCPUS); > > - pr_info("\t-i: Number of iterations per stage (default: %u)\n", > > - NR_TEST_ITERS_DEF); > > - pr_info("\t-p: Periodicity (in ms) of the guest timer (default: %u)\n", > > - TIMER_TEST_PERIOD_MS_DEF); > > - pr_info("\t-m: Frequency (in ms) of vCPUs to migrate to different pCPU. 0 to turn off (default: %u)\n", > > - TIMER_TEST_MIGRATION_FREQ_MS); > > - pr_info("\t-o: Counter offset (in counter cycles, default: 0)\n"); > > - pr_info("\t-h: print this help screen\n"); > > -} > > - > > -static bool parse_args(int argc, char *argv[]) > > -{ > > - int opt; > > - > > - while ((opt = getopt(argc, argv, "hn:i:p:m:o:")) != -1) { > > - switch (opt) { > > - case 'n': > > - test_args.nr_vcpus = atoi_positive("Number of vCPUs", optarg); > > - if (test_args.nr_vcpus > KVM_MAX_VCPUS) { > > - pr_info("Max allowed vCPUs: %u\n", > > - KVM_MAX_VCPUS); > > - goto err; > > - } > > - break; > > - case 'i': > > - test_args.nr_iter = atoi_positive("Number of iterations", optarg); > > - break; > > - case 'p': > > - test_args.timer_period_ms = atoi_positive("Periodicity", optarg); > > - break; > > - case 'm': > > - test_args.migration_freq_ms = atoi_non_negative("Frequency", optarg); > > - break; > > - case 'o': > > - test_args.offset.counter_offset = strtol(optarg, NULL, 0); > > - test_args.offset.reserved = 0; > > - break; > > - case 'h': > > - default: > > - goto err; > > - } > > - } > > - > > - return true; > > - > > -err: > > - test_print_help(argv[0]); > > - return false; > > -} > > - > > -int main(int argc, char *argv[]) > > -{ > > - struct kvm_vm *vm; > > - > > - if (!parse_args(argc, argv)) > > - exit(KSFT_SKIP); > > - > > - __TEST_REQUIRE(!test_args.migration_freq_ms || get_nprocs() >= 2, > > - "At least two physical CPUs needed for vCPU migration"); > > - > > - vm = test_vm_create(); > > - test_run(vm); > > - test_vm_cleanup(vm); > > - > > - return 0; > > -} > > diff --git a/tools/testing/selftests/kvm/arch_timer.c b/tools/testing/selftests/kvm/arch_timer.c > > new file mode 100644 > > index 000000000000..529024f58c98 > > --- /dev/null > > +++ b/tools/testing/selftests/kvm/arch_timer.c > > @@ -0,0 +1,252 @@ > > +// SPDX-License-Identifier: GPL-2.0-only > > +/* > > + * arch_timer.c - Tests the arch timer IRQ functionality > > + * > > + * The guest's main thread configures the timer interrupt for and waits > > It looks like the text was edited to remove the 'stage' references, which > is fine by me, but the 'for' should have also been removed. > > Sure! > > + * for it to fire, with a timeout equal to the timer period. > > + * It asserts that the timeout doesn't exceed the timer period. > > + * > > + * On the other hand, upon receipt of an interrupt, the guest's interrupt > > + * handler validates the interrupt by checking if the architectural state > > + * is in compliance with the specifications. > > + * > > + * The test provides command-line options to configure the timer's > > + * period (-p), number of vCPUs (-n), and iterations per stage (-i). > > + * To stress-test the timer stack even more, an option to migrate the > > + * vCPUs across pCPUs (-m), at a particular rate, is also provided. > > + * > > + * Copyright (c) 2021, Google LLC. > > + */ > > + > > +#define _GNU_SOURCE > > + > > +#include <stdlib.h> > > +#include <pthread.h> > > +#include <linux/sizes.h> > > +#include <linux/bitmap.h> > > +#include <sys/sysinfo.h> > > + > > +#include "timer_test.h" > > + > > +struct test_args test_args = { > > + .nr_vcpus = NR_VCPUS_DEF, > > + .nr_iter = NR_TEST_ITERS_DEF, > > + .timer_period_ms = TIMER_TEST_PERIOD_MS_DEF, > > + .migration_freq_ms = TIMER_TEST_MIGRATION_FREQ_MS, > > +#ifdef __aarch64__ > > + .offset = { .reserved = 1 }, > > +#endif > > Please run checkpatch, there are spaces instead of tabs in the struct. > Yes, the tabs were changed to spaces while copying. Will change it and run checkpatch to catch this kind of error in next version. > > +}; > > + > > +struct kvm_vcpu *vcpus[KVM_MAX_VCPUS]; > > +struct test_vcpu_shared_data vcpu_shared_data[KVM_MAX_VCPUS]; > > + > > +static pthread_t pt_vcpu_run[KVM_MAX_VCPUS]; > > +static unsigned long *vcpu_done_map; > > +static pthread_mutex_t vcpu_done_map_lock; > > + > > +static void *test_vcpu_run(void *arg) > > +{ > > + unsigned int vcpu_idx = (unsigned long)arg; > > + struct ucall uc; > > + struct kvm_vcpu *vcpu = vcpus[vcpu_idx]; > > + struct kvm_vm *vm = vcpu->vm; > > + struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx]; > > + > > + vcpu_run(vcpu); > > + > > + /* Currently, any exit from guest is an indication of completion */ > > + pthread_mutex_lock(&vcpu_done_map_lock); > > + __set_bit(vcpu_idx, vcpu_done_map); > > + pthread_mutex_unlock(&vcpu_done_map_lock); > > + > > + switch (get_ucall(vcpu, &uc)) { > > + case UCALL_SYNC: > > + case UCALL_DONE: > > + break; > > + case UCALL_ABORT: > > + sync_global_from_guest(vm, *shared_data); > > + fprintf(stderr, "Guest assert failed, vcpu %u; stage; %u; iter: %u\n", > > + vcpu_idx, shared_data->guest_stage, shared_data->nr_iter); > > + REPORT_GUEST_ASSERT(uc); > > + break; > > + default: > > + TEST_FAIL("Unexpected guest exit\n"); > > + } > > + > > + pr_info("PASS(vCPU-%d).\n", vcpu_idx); > > This is new. I can live with it, but generally we don't want to modify > functions while moving them. > Yes, this change was supposed to go with patch 8/8. > > + > > + return NULL; > > +} > > + > > +static uint32_t test_get_pcpu(void) > > +{ > > + uint32_t pcpu; > > + unsigned int nproc_conf; > > + cpu_set_t online_cpuset; > > + > > + nproc_conf = get_nprocs_conf(); > > + sched_getaffinity(0, sizeof(cpu_set_t), &online_cpuset); > > + > > + /* Randomly find an available pCPU to place a vCPU on */ > > + do { > > + pcpu = rand() % nproc_conf; > > + } while (!CPU_ISSET(pcpu, &online_cpuset)); > > + > > + return pcpu; > > +} > > + > > +static int test_migrate_vcpu(unsigned int vcpu_idx) > > +{ > > + int ret; > > + cpu_set_t cpuset; > > + uint32_t new_pcpu = test_get_pcpu(); > > + > > + CPU_ZERO(&cpuset); > > + CPU_SET(new_pcpu, &cpuset); > > + > > + pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu); > > + > > + ret = pthread_setaffinity_np(pt_vcpu_run[vcpu_idx], > > + sizeof(cpuset), &cpuset); > > + > > + /* Allow the error where the vCPU thread is already finished */ > > + TEST_ASSERT(ret == 0 || ret == ESRCH, > > + "Failed to migrate the vCPU:%u to pCPU: %u; ret: %d\n", > > + vcpu_idx, new_pcpu, ret); > > + > > + return ret; > > +} > > + > > +static void *test_vcpu_migration(void *arg) > > +{ > > + unsigned int i, n_done; > > + bool vcpu_done; > > + > > + do { > > + usleep(msecs_to_usecs(test_args.migration_freq_ms)); > > + > > + for (n_done = 0, i = 0; i < test_args.nr_vcpus; i++) { > > + pthread_mutex_lock(&vcpu_done_map_lock); > > + vcpu_done = test_bit(i, vcpu_done_map); > > + pthread_mutex_unlock(&vcpu_done_map_lock); > > + > > + if (vcpu_done) { > > + n_done++; > > + continue; > > + } > > + > > + test_migrate_vcpu(i); > > + } > > + } while (test_args.nr_vcpus != n_done); > > + > > + return NULL; > > +} > > + > > +static void test_run(struct kvm_vm *vm) > > +{ > > + pthread_t pt_vcpu_migration; > > + unsigned int i; > > + int ret; > > + > > + pthread_mutex_init(&vcpu_done_map_lock, NULL); > > + vcpu_done_map = bitmap_zalloc(test_args.nr_vcpus); > > + TEST_ASSERT(vcpu_done_map, "Failed to allocate vcpu done bitmap\n"); > > + > > + for (i = 0; i < (unsigned long)test_args.nr_vcpus; i++) { > > + ret = pthread_create(&pt_vcpu_run[i], NULL, test_vcpu_run, > > + (void *)(unsigned long)i); > > + TEST_ASSERT(!ret, "Failed to create vCPU-%d pthread\n", i); > > + } > > + > > + /* Spawn a thread to control the vCPU migrations */ > > + if (test_args.migration_freq_ms) { > > + srand(time(NULL)); > > + > > + ret = pthread_create(&pt_vcpu_migration, NULL, > > + test_vcpu_migration, NULL); > > + TEST_ASSERT(!ret, "Failed to create the migration pthread\n"); > > + } > > + > > + > > + for (i = 0; i < test_args.nr_vcpus; i++) > > + pthread_join(pt_vcpu_run[i], NULL); > > + > > + if (test_args.migration_freq_ms) > > + pthread_join(pt_vcpu_migration, NULL); > > + > > + bitmap_free(vcpu_done_map); > > +} > > + > > +static void test_print_help(char *name) > > +{ > > + pr_info("Usage: %s [-h] [-n nr_vcpus] [-i iterations] [-p timer_period_ms]\n", > > + name); > > + pr_info("\t-n: Number of vCPUs to configure (default: %u; max: %u)\n", > > + NR_VCPUS_DEF, KVM_MAX_VCPUS); > > + pr_info("\t-i: Number of iterations per stage (default: %u)\n", > > + NR_TEST_ITERS_DEF); > > + pr_info("\t-p: Periodicity (in ms) of the guest timer (default: %u)\n", > > + TIMER_TEST_PERIOD_MS_DEF); > > + pr_info("\t-m: Frequency (in ms) of vCPUs to migrate to different pCPU. 0 to turn off (default: %u)\n", > > + TIMER_TEST_MIGRATION_FREQ_MS); > > + pr_info("\t-o: Counter offset (in counter cycles, default: 0)\n"); > > + pr_info("\t-h: print this help screen\n"); > > +} > > + > > +static bool parse_args(int argc, char *argv[]) > > +{ > > + int opt; > > + > > + while ((opt = getopt(argc, argv, "hn:i:p:m:o:")) != -1) { > > + switch (opt) { > > + case 'n': > > + test_args.nr_vcpus = atoi_positive("Number of vCPUs", optarg); > > + if (test_args.nr_vcpus > KVM_MAX_VCPUS) { > > + pr_info("Max allowed vCPUs: %u\n", > > + KVM_MAX_VCPUS); > > + goto err; > > + } > > + break; > > + case 'i': > > + test_args.nr_iter = atoi_positive("Number of iterations", optarg); > > + break; > > + case 'p': > > + test_args.timer_period_ms = atoi_positive("Periodicity", optarg); > > + break; > > + case 'm': > > + test_args.migration_freq_ms = atoi_non_negative("Frequency", optarg); > > + break; > > + case 'o': > > + test_args.offset.counter_offset = strtol(optarg, NULL, 0); > > + test_args.offset.reserved = 0; > > + break; > > + case 'h': > > + default: > > + goto err; > > + } > > + } > > + > > + return true; > > + > > +err: > > + test_print_help(argv[0]); > > + return false; > > +} > > + > > +int main(int argc, char *argv[]) > > +{ > > + struct kvm_vm *vm; > > + > > + if (!parse_args(argc, argv)) > > + exit(KSFT_SKIP); > > + > > + __TEST_REQUIRE(!test_args.migration_freq_ms || get_nprocs() >= 2, > > + "At least two physical CPUs needed for vCPU migration"); > > + > > + vm = test_vm_create(); > > + test_run(vm); > > + test_vm_cleanup(vm); > > + > > + return 0; > > +} > > diff --git a/tools/testing/selftests/kvm/include/timer_test.h b/tools/testing/selftests/kvm/include/timer_test.h > > new file mode 100644 > > index 000000000000..109e4d635627 > > --- /dev/null > > +++ b/tools/testing/selftests/kvm/include/timer_test.h > > @@ -0,0 +1,52 @@ > > +/* SPDX-License-Identifier: GPL-2.0-only */ > > +/* > > + * tools/testing/selftests/kvm/include/timer_test.h > > + * > > + * Copyright (C) 2018, Google LLC > > + */ > > + > > +#ifndef SELFTEST_KVM_TIMER_TEST_H > > +#define SELFTEST_KVM_TIMER_TEST_H > > + > > +#include "kvm_util.h" > > + > > +#define NR_VCPUS_DEF 4 > > +#define NR_TEST_ITERS_DEF 5 > > +#define TIMER_TEST_PERIOD_MS_DEF 10 > > +#define TIMER_TEST_ERR_MARGIN_US 100 > > +#define TIMER_TEST_MIGRATION_FREQ_MS 2 > > + > > +#define msecs_to_usecs(msec) ((msec) * 1000LL) > > I'd move the above to include/test_util.h > Yes, msecs_to_usecs() macro should be a common API for all the tests. > > + > > +#define GICD_BASE_GPA 0x8000000ULL > > +#define GICR_BASE_GPA 0x80A0000ULL > > These defines belong in aarch64/arch_timer.c > These 2 defines were also defined in other test cases, shall we move them to an aarch64 specific header file? Maybe tools/testing/selftests/kvm/include/aarch64/gic.h? > > + > > +enum guest_stage { > > + GUEST_STAGE_VTIMER_CVAL=1, > > + GUEST_STAGE_VTIMER_TVAL, > > + GUEST_STAGE_PTIMER_CVAL, > > + GUEST_STAGE_PTIMER_TVAL, > > + GUEST_STAGE_MAX, > > +}; > > This enum also belongs in aarch64/arch_timer.c > Yes, it should be in aarch64/arch_timer.c > > + > > +/* Timer test cmdline parameters */ > > +struct test_args > > +{ > > + int nr_vcpus; > > + int nr_iter; > > + int timer_period_ms; > > + int migration_freq_ms; > > + struct kvm_arm_counter_offset offset; > > +}; > > + > > +/* Shared variables between host and guest */ > > +struct test_vcpu_shared_data { > > + int nr_iter; > > + enum guest_stage guest_stage; > > + uint64_t xcnt; > > +}; > > + > > +struct kvm_vm* test_vm_create(void); > > Move * next to function name. > Sure, thanks! > > +void test_vm_cleanup(struct kvm_vm *vm); > > + > > +#endif /* SELFTEST_KVM_TIMER_TEST_H */ > > -- > > 2.34.1 > > > > Thanks, > drew