[rh7,4/2,v2] kvm: use _safe version of list iteration in mmu_shrink_scan()

Submitted by Konstantin Khorenko on June 7, 2019, 12:56 p.m.

Details

Message ID 20190607125639.13754-1-khorenko@virtuozzo.com
State New
Series "Series without cover letter"
Headers show

Commit Message

Konstantin Khorenko June 7, 2019, 12:56 p.m.
As we skip some VMs during shrink and don't want to iterate them again
and again on each shrink, we move those skipped VMs to the list's tail,
thus we need to use _safe version of list iteration.

Fixes: bb2d7ab43eba ("kvm: move VMs which we skip during shrink to vm_list
tail")
https://jira.sw.ru/browse/PSBM-95077

Signed-off-by: Konstantin Khorenko <khorenko@virtuozzo.com>
---
 arch/x86/kvm/mmu.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

Patch hide | download patch | download mbox

diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 18c7f63fcccd..7d18cda1e2db 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -5343,13 +5343,13 @@  void kvm_mmu_invalidate_mmio_sptes(struct kvm *kvm, struct kvm_memslots *slots)
 static unsigned long
 mmu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
 {
-	struct kvm *kvm;
+	struct kvm *kvm, *tmp;
 	int nr_to_scan = sc->nr_to_scan;
 	unsigned long freed = 0;
 
 	spin_lock(&kvm_lock);
 
-	list_for_each_entry(kvm, &vm_list, vm_list) {
+	list_for_each_entry_safe(kvm, tmp, &vm_list, vm_list) {
 		int idx;
 		LIST_HEAD(invalid_list);