// SPDX-License-Identifier: GPL-2.0-only /* * KVM L1 hypervisor optimizations on Hyper-V. */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include #include #include "hyperv.h" #include "kvm_onhyperv.h" struct kvm_hv_tlb_range { u64 start_gfn; u64 pages; }; static int kvm_fill_hv_flush_list_func(struct hv_guest_mapping_flush_list *flush, void *data) { struct kvm_hv_tlb_range *range = data; return hyperv_fill_flush_guest_mapping_list(flush, range->start_gfn, range->pages); } static inline int hv_remote_flush_root_tdp(hpa_t root_tdp, struct kvm_hv_tlb_range *range) { if (range) return hyperv_flush_guest_mapping_range(root_tdp, kvm_fill_hv_flush_list_func, (void *)range); else return hyperv_flush_guest_mapping(root_tdp); } static int __hv_flush_remote_tlbs_range(struct kvm *kvm, struct kvm_hv_tlb_range *range) { struct kvm_arch *kvm_arch = &kvm->arch; struct kvm_vcpu *vcpu; int ret = 0, nr_unique_valid_roots; unsigned long i; hpa_t root; spin_lock(&kvm_arch->hv_root_tdp_lock); if (!VALID_PAGE(kvm_arch->hv_root_tdp)) { nr_unique_valid_roots = 0; /* * Flush all valid roots, and see if all vCPUs have converged * on a common root, in which case future flushes can skip the * loop and flush the common root. */ kvm_for_each_vcpu(i, vcpu, kvm) { root = vcpu->arch.hv_root_tdp; if (!VALID_PAGE(root) || root == kvm_arch->hv_root_tdp) continue; /* * Set the tracked root to the first valid root. Keep * this root for the entirety of the loop even if more * roots are encountered as a low effort optimization * to avoid flushing the same (first) root again. */ if (++nr_unique_valid_roots == 1) kvm_arch->hv_root_tdp = root; if (!ret) ret = hv_remote_flush_root_tdp(root, range); /* * Stop processing roots if a failure occurred and * multiple valid roots have already been detected. */ if (ret && nr_unique_valid_roots > 1) break; } /* * The optimized flush of a single root can't be used if there * are multiple valid roots (obviously). */ if (nr_unique_valid_roots > 1) kvm_arch->hv_root_tdp = INVALID_PAGE; } else { ret = hv_remote_flush_root_tdp(kvm_arch->hv_root_tdp, range); } spin_unlock(&kvm_arch->hv_root_tdp_lock); return ret; } int hv_flush_remote_tlbs_range(struct kvm *kvm, gfn_t start_gfn, gfn_t nr_pages) { struct kvm_hv_tlb_range range = { .start_gfn = start_gfn, .pages = nr_pages, }; return __hv_flush_remote_tlbs_range(kvm, &range); } EXPORT_SYMBOL_GPL(hv_flush_remote_tlbs_range); int hv_flush_remote_tlbs(struct kvm *kvm) { return __hv_flush_remote_tlbs_range(kvm, NULL); } EXPORT_SYMBOL_GPL(hv_flush_remote_tlbs); void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp) { struct kvm_arch *kvm_arch = &vcpu->kvm->arch; if (kvm_x86_ops.flush_remote_tlbs == hv_flush_remote_tlbs) { spin_lock(&kvm_arch->hv_root_tdp_lock); vcpu->arch.hv_root_tdp = root_tdp; if (root_tdp != kvm_arch->hv_root_tdp) kvm_arch->hv_root_tdp = INVALID_PAGE; spin_unlock(&kvm_arch->hv_root_tdp_lock); } } EXPORT_SYMBOL_GPL(hv_track_root_tdp);