/* SPDX-License-Identifier: GPL-2.0 */ /* * KVM guest address space mapping code * * Copyright IBM Corp. 2007, 2016, 2025 * Author(s): Martin Schwidefsky * Claudio Imbrenda */ #ifndef ARCH_KVM_S390_GMAP_H #define ARCH_KVM_S390_GMAP_H #include "dat.h" /** * enum gmap_flags - Flags of a gmap. * * @GMAP_FLAG_SHADOW: The gmap is a vsie shadow gmap. * @GMAP_FLAG_OWNS_PAGETABLES: The gmap owns all dat levels; normally 1, is 0 * only for ucontrol per-cpu gmaps, since they * share the page tables with the main gmap. * @GMAP_FLAG_IS_UCONTROL: The gmap is ucontrol (main gmap or per-cpu gmap). * @GMAP_FLAG_ALLOW_HPAGE_1M: 1M hugepages are allowed for this gmap, * independently of the page size used by userspace. * @GMAP_FLAG_ALLOW_HPAGE_2G: 2G hugepages are allowed for this gmap, * independently of the page size used by userspace. * @GMAP_FLAG_PFAULT_ENABLED: Pfault is enabled for the gmap. * @GMAP_FLAG_USES_SKEYS: If the guest uses storage keys. * @GMAP_FLAG_USES_CMM: Whether the guest uses CMMA. * @GMAP_FLAG_EXPORT_ON_UNMAP: Whether to export guest pages when unmapping. */ enum gmap_flags { GMAP_FLAG_SHADOW = 0, GMAP_FLAG_OWNS_PAGETABLES, GMAP_FLAG_IS_UCONTROL, GMAP_FLAG_ALLOW_HPAGE_1M, GMAP_FLAG_ALLOW_HPAGE_2G, GMAP_FLAG_PFAULT_ENABLED, GMAP_FLAG_USES_SKEYS, GMAP_FLAG_USES_CMM, GMAP_FLAG_EXPORT_ON_UNMAP, }; /** * struct gmap_struct - Guest address space. * * @flags: GMAP_FLAG_* flags. * @edat_level: The edat level of this shadow gmap. * @kvm: The vm. * @asce: The ASCE used by this gmap. * @list: List head used in children gmaps for the children gmap list. * @children_lock: Protects children and scb_users. * @children: List of child gmaps of this gmap. * @scb_users: List of vsie_scb that use this shadow gmap. * @parent: Parent gmap of a child gmap. * @guest_asce: Original ASCE of this shadow gmap. * @host_to_rmap_lock: Protects host_to_rmap. * @host_to_rmap: Radix tree mapping host addresses to guest addresses. */ struct gmap { unsigned long flags; unsigned char edat_level; struct kvm *kvm; union asce asce; struct list_head list; spinlock_t children_lock; /* Protects: children, scb_users */ struct list_head children; struct list_head scb_users; struct gmap *parent; union asce guest_asce; spinlock_t host_to_rmap_lock; /* Protects host_to_rmap */ struct radix_tree_root host_to_rmap; refcount_t refcount; }; struct gmap_cache { struct list_head list; struct gmap *gmap; }; #define gmap_for_each_rmap_safe(pos, n, head) \ for (pos = (head); n = pos ? pos->next : NULL, pos; pos = n) int s390_replace_asce(struct gmap *gmap); bool _gmap_unmap_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end, bool hint); bool gmap_age_gfn(struct gmap *gmap, gfn_t start, gfn_t end); bool gmap_unmap_gfn_range(struct gmap *gmap, struct kvm_memory_slot *slot, gfn_t start, gfn_t end); int gmap_try_fixup_minor(struct gmap *gmap, struct guest_fault *fault); struct gmap *gmap_new(struct kvm *kvm, gfn_t limit); struct gmap *gmap_new_child(struct gmap *parent, gfn_t limit); void gmap_remove_child(struct gmap *child); void gmap_dispose(struct gmap *gmap); int gmap_link(struct kvm_s390_mmu_cache *mc, struct gmap *gmap, struct guest_fault *fault); void gmap_sync_dirty_log(struct gmap *gmap, gfn_t start, gfn_t end); int gmap_set_limit(struct gmap *gmap, gfn_t limit); int gmap_ucas_translate(struct kvm_s390_mmu_cache *mc, struct gmap *gmap, gpa_t *gaddr); int gmap_ucas_map(struct gmap *gmap, gfn_t p_gfn, gfn_t c_gfn, unsigned long count); void gmap_ucas_unmap(struct gmap *gmap, gfn_t c_gfn, unsigned long count); int gmap_enable_skeys(struct gmap *gmap); int gmap_pv_destroy_range(struct gmap *gmap, gfn_t start, gfn_t end, bool interruptible); int gmap_insert_rmap(struct gmap *sg, gfn_t p_gfn, gfn_t r_gfn, int level); int gmap_protect_rmap(struct kvm_s390_mmu_cache *mc, struct gmap *sg, gfn_t p_gfn, gfn_t r_gfn, kvm_pfn_t pfn, int level, bool wr); void gmap_set_cmma_all_dirty(struct gmap *gmap); void _gmap_handle_vsie_unshadow_event(struct gmap *parent, gfn_t gfn); struct gmap *gmap_create_shadow(struct kvm_s390_mmu_cache *mc, struct gmap *gmap, union asce asce, int edat_level); void gmap_split_huge_pages(struct gmap *gmap); static inline bool uses_skeys(struct gmap *gmap) { return test_bit(GMAP_FLAG_USES_SKEYS, &gmap->flags); } static inline bool uses_cmm(struct gmap *gmap) { return test_bit(GMAP_FLAG_USES_CMM, &gmap->flags); } static inline bool pfault_enabled(struct gmap *gmap) { return test_bit(GMAP_FLAG_PFAULT_ENABLED, &gmap->flags); } static inline bool is_ucontrol(struct gmap *gmap) { return test_bit(GMAP_FLAG_IS_UCONTROL, &gmap->flags); } static inline bool is_shadow(struct gmap *gmap) { return test_bit(GMAP_FLAG_SHADOW, &gmap->flags); } static inline bool owns_page_tables(struct gmap *gmap) { return test_bit(GMAP_FLAG_OWNS_PAGETABLES, &gmap->flags); } static inline struct gmap *gmap_put(struct gmap *gmap) { if (refcount_dec_and_test(&gmap->refcount)) gmap_dispose(gmap); return NULL; } static inline void gmap_get(struct gmap *gmap) { WARN_ON_ONCE(unlikely(!refcount_inc_not_zero(&gmap->refcount))); } static inline void gmap_handle_vsie_unshadow_event(struct gmap *parent, gfn_t gfn) { scoped_guard(spinlock, &parent->children_lock) _gmap_handle_vsie_unshadow_event(parent, gfn); } static inline bool gmap_mkold_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end) { return _gmap_unmap_prefix(gmap, gfn, end, true); } static inline bool gmap_unmap_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end) { return _gmap_unmap_prefix(gmap, gfn, end, false); } static inline union pgste _gmap_ptep_xchg(struct gmap *gmap, union pte *ptep, union pte newpte, union pgste pgste, gfn_t gfn, bool needs_lock) { lockdep_assert_held(&gmap->kvm->mmu_lock); if (!needs_lock) lockdep_assert_held(&gmap->children_lock); else lockdep_assert_not_held(&gmap->children_lock); if (pgste.prefix_notif && (newpte.h.p || newpte.h.i)) { pgste.prefix_notif = 0; gmap_unmap_prefix(gmap, gfn, gfn + 1); } if (pgste.vsie_notif && (ptep->h.p != newpte.h.p || newpte.h.i)) { pgste.vsie_notif = 0; if (needs_lock) gmap_handle_vsie_unshadow_event(gmap, gfn); else _gmap_handle_vsie_unshadow_event(gmap, gfn); } return __dat_ptep_xchg(ptep, pgste, newpte, gfn, gmap->asce, uses_skeys(gmap)); } static inline union pgste gmap_ptep_xchg(struct gmap *gmap, union pte *ptep, union pte newpte, union pgste pgste, gfn_t gfn) { return _gmap_ptep_xchg(gmap, ptep, newpte, pgste, gfn, true); } static inline void _gmap_crstep_xchg(struct gmap *gmap, union crste *crstep, union crste ne, gfn_t gfn, bool needs_lock) { unsigned long align = 8 + (is_pmd(*crstep) ? 0 : 11); lockdep_assert_held(&gmap->kvm->mmu_lock); if (!needs_lock) lockdep_assert_held(&gmap->children_lock); gfn = ALIGN_DOWN(gfn, align); if (crste_prefix(*crstep) && (ne.h.p || ne.h.i || !crste_prefix(ne))) { ne.s.fc1.prefix_notif = 0; gmap_unmap_prefix(gmap, gfn, gfn + align); } if (crste_leaf(*crstep) && crstep->s.fc1.vsie_notif && (ne.h.p || ne.h.i || !ne.s.fc1.vsie_notif)) { ne.s.fc1.vsie_notif = 0; if (needs_lock) gmap_handle_vsie_unshadow_event(gmap, gfn); else _gmap_handle_vsie_unshadow_event(gmap, gfn); } dat_crstep_xchg(crstep, ne, gfn, gmap->asce); } static inline void gmap_crstep_xchg(struct gmap *gmap, union crste *crstep, union crste ne, gfn_t gfn) { return _gmap_crstep_xchg(gmap, crstep, ne, gfn, true); } /** * gmap_is_shadow_valid() - check if a shadow guest address space matches the * given properties and is still valid. * @sg: Pointer to the shadow guest address space structure. * @asce: ASCE for which the shadow table is requested. * @edat_level: Edat level to be used for the shadow translation. * * Return: true if the gmap shadow is still valid and matches the given * properties and the caller can continue using it; false otherwise, the * caller has to request a new shadow gmap in this case. */ static inline bool gmap_is_shadow_valid(struct gmap *sg, union asce asce, int edat_level) { return sg->guest_asce.val == asce.val && sg->edat_level == edat_level; } #endif /* ARCH_KVM_S390_GMAP_H */