summaryrefslogtreecommitdiff
path: root/arch/s390/kvm/faultin.h
blob: f86176d2769c5fe8ff41a9338b2aa90f5b9b924f (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
/* SPDX-License-Identifier: GPL-2.0 */
/*
 *  KVM guest fault handling.
 *
 *    Copyright IBM Corp. 2025
 *    Author(s): Claudio Imbrenda <imbrenda@linux.ibm.com>
 */

#ifndef __KVM_S390_FAULTIN_H
#define __KVM_S390_FAULTIN_H

#include <linux/kvm_host.h>

#include "dat.h"

int kvm_s390_faultin_gfn(struct kvm_vcpu *vcpu, struct kvm *kvm, struct guest_fault *f);
int kvm_s390_get_guest_page(struct kvm *kvm, struct guest_fault *f, gfn_t gfn, bool w);

static inline int kvm_s390_faultin_gfn_simple(struct kvm_vcpu *vcpu, struct kvm *kvm,
					      gfn_t gfn, bool wr)
{
	struct guest_fault f = { .gfn = gfn, .write_attempt = wr, };

	return kvm_s390_faultin_gfn(vcpu, kvm, &f);
}

static inline int kvm_s390_get_guest_page_and_read_gpa(struct kvm *kvm, struct guest_fault *f,
						       gpa_t gaddr, unsigned long *val)
{
	int rc;

	rc = kvm_s390_get_guest_page(kvm, f, gpa_to_gfn(gaddr), false);
	if (rc)
		return rc;

	*val = *(unsigned long *)phys_to_virt(pfn_to_phys(f->pfn) | offset_in_page(gaddr));

	return 0;
}

static inline void kvm_s390_release_multiple(struct kvm *kvm, struct guest_fault *guest_faults,
					     int n, bool ignore)
{
	int i;

	for (i = 0; i < n; i++) {
		kvm_release_faultin_page(kvm, guest_faults[i].page, ignore,
					 guest_faults[i].write_attempt);
		guest_faults[i].page = NULL;
	}
}

static inline bool kvm_s390_multiple_faults_need_retry(struct kvm *kvm, unsigned long seq,
						       struct guest_fault *guest_faults, int n,
						       bool unsafe)
{
	int i;

	for (i = 0; i < n; i++) {
		if (!guest_faults[i].valid)
			continue;
		if (unsafe && mmu_invalidate_retry_gfn_unsafe(kvm, seq, guest_faults[i].gfn))
			return true;
		if (!unsafe && mmu_invalidate_retry_gfn(kvm, seq, guest_faults[i].gfn))
			return true;
	}
	return false;
}

static inline int kvm_s390_get_guest_pages(struct kvm *kvm, struct guest_fault *guest_faults,
					   gfn_t start, int n_pages, bool write_attempt)
{
	int i, rc;

	for (i = 0; i < n_pages; i++) {
		rc = kvm_s390_get_guest_page(kvm, guest_faults + i, start + i, write_attempt);
		if (rc)
			break;
	}
	return rc;
}

#define kvm_s390_release_faultin_array(kvm, array, ignore) \
	kvm_s390_release_multiple(kvm, array, ARRAY_SIZE(array), ignore)

#define kvm_s390_array_needs_retry_unsafe(kvm, seq, array) \
	kvm_s390_multiple_faults_need_retry(kvm, seq, array, ARRAY_SIZE(array), true)

#define kvm_s390_array_needs_retry_safe(kvm, seq, array) \
	kvm_s390_multiple_faults_need_retry(kvm, seq, array, ARRAY_SIZE(array), false)

#endif /* __KVM_S390_FAULTIN_H */