// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2019 Western Digital Corporation or its affiliates. * * Authors: * Anup Patel */ #include #include #include #include #include #include #include #include static unsigned long vmid_version = 1; static unsigned long vmid_next; static unsigned long vmid_bits; static DEFINE_SPINLOCK(vmid_lock); void kvm_riscv_stage2_vmid_detect(void) { unsigned long old; /* Figure-out number of VMID bits in HW */ old = csr_read(CSR_HGATP); csr_write(CSR_HGATP, old | HGATP_VMID_MASK); vmid_bits = csr_read(CSR_HGATP); vmid_bits = (vmid_bits & HGATP_VMID_MASK) >> HGATP_VMID_SHIFT; vmid_bits = fls_long(vmid_bits); csr_write(CSR_HGATP, old); /* We polluted local TLB so flush all guest TLB */ __kvm_riscv_hfence_gvma_all(); /* We don't use VMID bits if they are not sufficient */ if ((1UL << vmid_bits) < num_possible_cpus()) vmid_bits = 0; } unsigned long kvm_riscv_stage2_vmid_bits(void) { return vmid_bits; } int kvm_riscv_stage2_vmid_init(struct kvm *kvm) { /* Mark the initial VMID and VMID version invalid */ kvm->arch.vmid.vmid_version = 0; kvm->arch.vmid.vmid = 0; return 0; } bool kvm_riscv_stage2_vmid_ver_changed(struct kvm_vmid *vmid) { if (!vmid_bits) return false; return unlikely(READ_ONCE(vmid->vmid_version) != READ_ONCE(vmid_version)); } void kvm_riscv_stage2_vmid_update(struct kvm_vcpu *vcpu) { int i; struct kvm_vcpu *v; struct cpumask hmask; struct kvm_vmid *vmid = &vcpu->kvm->arch.vmid; if (!kvm_riscv_stage2_vmid_ver_changed(vmid)) return; spin_lock(&vmid_lock); /* * We need to re-check the vmid_version here to ensure that if * another vcpu already allocated a valid vmid for this vm. */ if (!kvm_riscv_stage2_vmid_ver_changed(vmid)) { spin_unlock(&vmid_lock); return; } /* First user of a new VMID version? */ if (unlikely(vmid_next == 0)) { WRITE_ONCE(vmid_version, READ_ONCE(vmid_version) + 1); vmid_next = 1; /* * We ran out of VMIDs so we increment vmid_version and * start assigning VMIDs from 1. * * This also means existing VMIDs assignement to all Guest * instances is invalid and we have force VMID re-assignement * for all Guest instances. The Guest instances that were not * running will automatically pick-up new VMIDs because will * call kvm_riscv_stage2_vmid_update() whenever they enter * in-kernel run loop. For Guest instances that are already * running, we force VM exits on all host CPUs using IPI and * flush all Guest TLBs. */ riscv_cpuid_to_hartid_mask(cpu_online_mask, &hmask); sbi_remote_hfence_gvma(cpumask_bits(&hmask), 0, 0); } vmid->vmid = vmid_next; vmid_next++; vmid_next &= (1 << vmid_bits) - 1; WRITE_ONCE(vmid->vmid_version, READ_ONCE(vmid_version)); spin_unlock(&vmid_lock); /* Request stage2 page table update for all VCPUs */ kvm_for_each_vcpu(i, v, vcpu->kvm) kvm_make_request(KVM_REQ_UPDATE_HGATP, v); }