OpenWrt – Rev 1

Subversion Repositories:
Rev:
From 621f2c4753b03170213e178cdafd66e78b212b3c Mon Sep 17 00:00:00 2001
From: Biwen Li <biwen.li@nxp.com>
Date: Tue, 30 Oct 2018 18:26:41 +0800
Subject: [PATCH 27/40] kvm: support layerscape
This is an integrated patch of kvm for layerscape

Signed-off-by: Bharat Bhushan <Bharat.Bhushan@nxp.com>
Signed-off-by: Laurentiu Tudor <laurentiu.tudor@nxp.com>
Signed-off-by: Signed-off-by: Biwen Li <biwen.li@nxp.com>
---
 arch/arm/include/asm/kvm_mmu.h   |  3 +-
 arch/arm64/include/asm/kvm_mmu.h | 14 ++++++--
 arch/powerpc/kvm/booke.c         |  5 +++
 virt/kvm/arm/mmu.c               | 56 ++++++++++++++++++++++++++++++--
 virt/kvm/arm/vgic/vgic-its.c     |  2 +-
 virt/kvm/arm/vgic/vgic-v2.c      |  3 +-
 6 files changed, 74 insertions(+), 9 deletions(-)

--- a/arch/arm/include/asm/kvm_mmu.h
+++ b/arch/arm/include/asm/kvm_mmu.h
@@ -55,7 +55,8 @@ void stage2_unmap_vm(struct kvm *kvm);
 int kvm_alloc_stage2_pgd(struct kvm *kvm);
 void kvm_free_stage2_pgd(struct kvm *kvm);
 int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
-                         phys_addr_t pa, unsigned long size, bool writable);
+                         phys_addr_t pa, unsigned long size, bool writable,
+                         pgprot_t prot);
 
 int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
 
--- a/arch/arm64/include/asm/kvm_mmu.h
+++ b/arch/arm64/include/asm/kvm_mmu.h
@@ -167,7 +167,8 @@ void stage2_unmap_vm(struct kvm *kvm);
 int kvm_alloc_stage2_pgd(struct kvm *kvm);
 void kvm_free_stage2_pgd(struct kvm *kvm);
 int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
-                         phys_addr_t pa, unsigned long size, bool writable);
+                         phys_addr_t pa, unsigned long size, bool writable,
+                         pgprot_t prot);
 
 int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
 
@@ -270,8 +271,15 @@ static inline void __coherent_cache_gues
 
 static inline void __kvm_flush_dcache_pte(pte_t pte)
 {
-       struct page *page = pte_page(pte);
-       kvm_flush_dcache_to_poc(page_address(page), PAGE_SIZE);
+       if (pfn_valid(pte_pfn(pte))) {
+               struct page *page = pte_page(pte);
+               kvm_flush_dcache_to_poc(page_address(page), PAGE_SIZE);
+       } else {
+               void __iomem *va = ioremap_cache_ns(pte_pfn(pte) << PAGE_SHIFT, PAGE_SIZE);
+
+               kvm_flush_dcache_to_poc(va, PAGE_SIZE);
+               iounmap(va);
+       }
 }
 
 static inline void __kvm_flush_dcache_pmd(pmd_t pmd)
--- a/arch/powerpc/kvm/booke.c
+++ b/arch/powerpc/kvm/booke.c
@@ -305,6 +305,11 @@ void kvmppc_core_queue_fpunavail(struct
        kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_FP_UNAVAIL);
 }
 
+void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu)
+{
+       kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_ALTIVEC_UNAVAIL);
+}
+
 void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu)
 {
        kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_DECREMENTER);
--- a/virt/kvm/arm/mmu.c
+++ b/virt/kvm/arm/mmu.c
@@ -1027,9 +1027,11 @@ static int stage2_pmdp_test_and_clear_yo
  * @guest_ipa: The IPA at which to insert the mapping
  * @pa:                The physical address of the device
  * @size:      The size of the mapping
+ * @prot:      S2 page translation bits
  */
 int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
-                         phys_addr_t pa, unsigned long size, bool writable)
+                         phys_addr_t pa, unsigned long size, bool writable,
+                         pgprot_t prot)
 {
        phys_addr_t addr, end;
        int ret = 0;
@@ -1040,7 +1042,7 @@ int kvm_phys_addr_ioremap(struct kvm *kv
        pfn = __phys_to_pfn(pa);
 
        for (addr = guest_ipa; addr < end; addr += PAGE_SIZE) {
-               pte_t pte = pfn_pte(pfn, PAGE_S2_DEVICE);
+               pte_t pte = pfn_pte(pfn, prot);
 
                if (writable)
                        pte = kvm_s2pte_mkwrite(pte);
@@ -1064,6 +1066,30 @@ out:
        return ret;
 }
 
+#ifdef CONFIG_ARM64
+static pgprot_t stage1_to_stage2_pgprot(pgprot_t prot)
+{
+       switch (pgprot_val(prot) & PTE_ATTRINDX_MASK) {
+               case PTE_ATTRINDX(MT_DEVICE_nGnRE):
+               case PTE_ATTRINDX(MT_DEVICE_nGnRnE):
+               case PTE_ATTRINDX(MT_DEVICE_GRE):
+                       return PAGE_S2_DEVICE;
+               case PTE_ATTRINDX(MT_NORMAL_NC):
+               case PTE_ATTRINDX(MT_NORMAL):
+                       return (pgprot_val(prot) & PTE_SHARED)
+                               ? PAGE_S2
+                               : PAGE_S2_NS;
+       }
+
+       return PAGE_S2_DEVICE;
+}
+#else
+static pgprot_t stage1_to_stage2_pgprot(pgprot_t prot)
+{
+       return PAGE_S2_DEVICE;
+}
+#endif
+
 static bool transparent_hugepage_adjust(kvm_pfn_t *pfnp, phys_addr_t *ipap)
 {
        kvm_pfn_t pfn = *pfnp;
@@ -1334,6 +1360,18 @@ static int user_mem_abort(struct kvm_vcp
                hugetlb = true;
                gfn = (fault_ipa & PMD_MASK) >> PAGE_SHIFT;
        } else {
+               pte_t *pte;
+               spinlock_t *ptl;
+               pgprot_t prot;
+
+               pte = get_locked_pte(current->mm, memslot->userspace_addr, &ptl);
+               prot = stage1_to_stage2_pgprot(__pgprot(pte_val(*pte)));
+               pte_unmap_unlock(pte, ptl);
+#ifdef CONFIG_ARM64
+               if (pgprot_val(prot) == pgprot_val(PAGE_S2_NS))
+                       mem_type = PAGE_S2_NS;
+#endif
+
                /*
                 * Pages belonging to memslots that don't have the same
                 * alignment for userspace and IPA cannot be mapped using
@@ -1375,6 +1413,11 @@ static int user_mem_abort(struct kvm_vcp
        if (is_error_noslot_pfn(pfn))
                return -EFAULT;
 
+#ifdef CONFIG_ARM64
+       if (pgprot_val(mem_type) == pgprot_val(PAGE_S2_NS)) {
+               flags |= KVM_S2PTE_FLAG_IS_IOMAP;
+       } else
+#endif
        if (kvm_is_device_pfn(pfn)) {
                mem_type = PAGE_S2_DEVICE;
                flags |= KVM_S2PTE_FLAG_IS_IOMAP;
@@ -1911,6 +1954,9 @@ int kvm_arch_prepare_memory_region(struc
                        gpa_t gpa = mem->guest_phys_addr +
                                    (vm_start - mem->userspace_addr);
                        phys_addr_t pa;
+                       pgprot_t prot;
+                       pte_t *pte;
+                       spinlock_t *ptl;
 
                        pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT;
                        pa += vm_start - vma->vm_start;
@@ -1921,9 +1967,13 @@ int kvm_arch_prepare_memory_region(struc
                                goto out;
                        }
 
+                       pte = get_locked_pte(current->mm, mem->userspace_addr, &ptl);
+                       prot = stage1_to_stage2_pgprot(__pgprot(pte_val(*pte)));
+                       pte_unmap_unlock(pte, ptl);
+
                        ret = kvm_phys_addr_ioremap(kvm, gpa, pa,
                                                    vm_end - vm_start,
-                                                   writable);
+                                                   writable, prot);
                        if (ret)
                                break;
                }
--- a/virt/kvm/arm/vgic/vgic-its.c
+++ b/virt/kvm/arm/vgic/vgic-its.c
@@ -243,7 +243,7 @@ static struct its_ite *find_ite(struct v
 #define GIC_LPI_OFFSET 8192
 
 #define VITS_TYPER_IDBITS 16
-#define VITS_TYPER_DEVBITS 16
+#define VITS_TYPER_DEVBITS 17
 #define VITS_DTE_MAX_DEVID_OFFSET      (BIT(14) - 1)
 #define VITS_ITE_MAX_EVENTID_OFFSET    (BIT(16) - 1)
 
--- a/virt/kvm/arm/vgic/vgic-v2.c
+++ b/virt/kvm/arm/vgic/vgic-v2.c
@@ -304,7 +304,8 @@ int vgic_v2_map_resources(struct kvm *kv
        if (!static_branch_unlikely(&vgic_v2_cpuif_trap)) {
                ret = kvm_phys_addr_ioremap(kvm, dist->vgic_cpu_base,
                                            kvm_vgic_global_state.vcpu_base,
-                                           KVM_VGIC_V2_CPU_SIZE, true);
+                                           KVM_VGIC_V2_CPU_SIZE, true,
+                                           PAGE_S2_DEVICE);
                if (ret) {
                        kvm_err("Unable to remap VGIC CPU to VCPU\n");
                        goto out;