OpenWrt – Rev 2

Subversion Repositories:
Rev:
From 48dbe4b3a31795b8efdfff82f69eccd086052eed Mon Sep 17 00:00:00 2001
From: Biwen Li <biwen.li@nxp.com>
Date: Fri, 16 Nov 2018 10:27:30 +0800
Subject: [PATCH 16/39] dpaa-bqman: support layerscape 
This is an integrated patch of dpaa-bqman for layerscape

Signed-off-by: Claudiu Manoil <claudiu.manoil@nxp.com>
Signed-off-by: Madalin Bucur <madalin.bucur@nxp.com>
Signed-off-by: Roy Pledge <roy.pledge@nxp.com>
Signed-off-by: Stuart Yoder <stuart.yoder@nxp.com>
Signed-off-by: Valentin Rothberg <valentinrothberg@gmail.com>
Signed-off-by: Biwen Li <biwen.li@nxp.com>
---
 drivers/soc/fsl/qbman/Kconfig       |   2 +-
 drivers/soc/fsl/qbman/bman.c        |  24 ++++-
 drivers/soc/fsl/qbman/bman_ccsr.c   |  35 ++++++-
 drivers/soc/fsl/qbman/bman_portal.c |  12 ++-
 drivers/soc/fsl/qbman/bman_priv.h   |   3 +
 drivers/soc/fsl/qbman/dpaa_sys.h    |   8 +-
 drivers/soc/fsl/qbman/qman.c        |  46 ++++++++-
 drivers/soc/fsl/qbman/qman_ccsr.c   | 140 ++++++++++++++++++++++------
 drivers/soc/fsl/qbman/qman_portal.c |  12 ++-
 drivers/soc/fsl/qbman/qman_priv.h   |   5 +-
 drivers/soc/fsl/qbman/qman_test.h   |   2 -
 11 files changed, 236 insertions(+), 53 deletions(-)

--- a/drivers/soc/fsl/qbman/Kconfig
+++ b/drivers/soc/fsl/qbman/Kconfig
@@ -1,6 +1,6 @@
 menuconfig FSL_DPAA
        bool "Freescale DPAA 1.x support"
-       depends on FSL_SOC_BOOKE
+       depends on (FSL_SOC_BOOKE || ARCH_LAYERSCAPE)
        select GENERIC_ALLOCATOR
        help
          The Freescale Data Path Acceleration Architecture (DPAA) is a set of
--- a/drivers/soc/fsl/qbman/bman.c
+++ b/drivers/soc/fsl/qbman/bman.c
@@ -35,6 +35,27 @@
 
 /* Portal register assists */
 
+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
+/* Cache-inhibited register offsets */
+#define BM_REG_RCR_PI_CINH     0x3000
+#define BM_REG_RCR_CI_CINH     0x3100
+#define BM_REG_RCR_ITR         0x3200
+#define BM_REG_CFG             0x3300
+#define BM_REG_SCN(n)          (0x3400 + ((n) << 6))
+#define BM_REG_ISR             0x3e00
+#define BM_REG_IER             0x3e40
+#define BM_REG_ISDR            0x3e80
+#define BM_REG_IIR             0x3ec0
+
+/* Cache-enabled register offsets */
+#define BM_CL_CR               0x0000
+#define BM_CL_RR0              0x0100
+#define BM_CL_RR1              0x0140
+#define BM_CL_RCR              0x1000
+#define BM_CL_RCR_PI_CENA      0x3000
+#define BM_CL_RCR_CI_CENA      0x3100
+
+#else
 /* Cache-inhibited register offsets */
 #define BM_REG_RCR_PI_CINH     0x0000
 #define BM_REG_RCR_CI_CINH     0x0004
@@ -53,6 +74,7 @@
 #define BM_CL_RCR              0x1000
 #define BM_CL_RCR_PI_CENA      0x3000
 #define BM_CL_RCR_CI_CENA      0x3100
+#endif
 
 /*
  * Portal modes.
@@ -607,7 +629,7 @@ int bman_p_irqsource_add(struct bman_por
        unsigned long irqflags;
 
        local_irq_save(irqflags);
-       set_bits(bits & BM_PIRQ_VISIBLE, &p->irq_sources);
+       p->irq_sources |= bits & BM_PIRQ_VISIBLE;
        bm_out(&p->p, BM_REG_IER, p->irq_sources);
        local_irq_restore(irqflags);
        return 0;
--- a/drivers/soc/fsl/qbman/bman_ccsr.c
+++ b/drivers/soc/fsl/qbman/bman_ccsr.c
@@ -170,10 +170,11 @@ static int fsl_bman_probe(struct platfor
 {
        int ret, err_irq;
        struct device *dev = &pdev->dev;
-       struct device_node *node = dev->of_node;
+       struct device_node *mem_node, *node = dev->of_node;
        struct resource *res;
        u16 id, bm_pool_cnt;
        u8 major, minor;
+       u64 size;
 
        res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
        if (!res) {
@@ -201,6 +202,38 @@ static int fsl_bman_probe(struct platfor
                return -ENODEV;
        }
 
+       /*
+        * If FBPR memory wasn't defined using the qbman compatiable string
+        * try using the of_reserved_mem_device method
+        */
+       if (!fbpr_a) {
+               ret = of_reserved_mem_device_init(dev);
+               if (ret) {
+                       dev_err(dev, "of_reserved_mem_device_init() failed 0x%x\n",
+                               ret);
+                       return -ENODEV;
+               }
+               mem_node = of_parse_phandle(dev->of_node, "memory-region", 0);
+               if (mem_node) {
+                       ret = of_property_read_u64(mem_node, "size", &size);
+                       if (ret) {
+                               dev_err(dev, "FBPR: of_address_to_resource fails 0x%x\n",
+                                       ret);
+                               return -ENODEV;
+                       }
+                       fbpr_sz = size;
+               } else {
+                       dev_err(dev, "No memory-region found for FBPR\n");
+                       return -ENODEV;
+               }
+               if (!dma_zalloc_coherent(dev, fbpr_sz, &fbpr_a, 0)) {
+                       dev_err(dev, "Alloc FBPR memory failed\n");
+                       return -ENODEV;
+               }
+       }
+
+       dev_dbg(dev, "Allocated FBPR 0x%llx 0x%zx\n", fbpr_a, fbpr_sz);
+
        bm_set_memory(fbpr_a, fbpr_sz);
 
        err_irq = platform_get_irq(pdev, 0);
--- a/drivers/soc/fsl/qbman/bman_portal.c
+++ b/drivers/soc/fsl/qbman/bman_portal.c
@@ -123,7 +123,14 @@ static int bman_portal_probe(struct plat
        }
        pcfg->irq = irq;
 
-       va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]), 0);
+#ifdef CONFIG_PPC
+       /* PPC requires a cacheable/non-coherent mapping of the portal */
+       va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]),
+                         (pgprot_val(PAGE_KERNEL) & ~_PAGE_COHERENT));
+#else
+       /* For ARM we can use write combine mapping. */
+       va = ioremap_wc(addr_phys[0]->start, resource_size(addr_phys[0]));
+#endif
        if (!va) {
                dev_err(dev, "ioremap::CE failed\n");
                goto err_ioremap1;
@@ -131,8 +138,7 @@ static int bman_portal_probe(struct plat
 
        pcfg->addr_virt[DPAA_PORTAL_CE] = va;
 
-       va = ioremap_prot(addr_phys[1]->start, resource_size(addr_phys[1]),
-                         _PAGE_GUARDED | _PAGE_NO_CACHE);
+       va = ioremap(addr_phys[1]->start, resource_size(addr_phys[1]));
        if (!va) {
                dev_err(dev, "ioremap::CI failed\n");
                goto err_ioremap2;
--- a/drivers/soc/fsl/qbman/bman_priv.h
+++ b/drivers/soc/fsl/qbman/bman_priv.h
@@ -33,6 +33,9 @@
 #include "dpaa_sys.h"
 
 #include <soc/fsl/bman.h>
+#include <linux/dma-contiguous.h>
+#include <linux/of_address.h>
+#include <linux/dma-mapping.h>
 
 /* Portal processing (interrupt) sources */
 #define BM_PIRQ_RCRI   0x00000002      /* RCR Ring (below threshold) */
--- a/drivers/soc/fsl/qbman/dpaa_sys.h
+++ b/drivers/soc/fsl/qbman/dpaa_sys.h
@@ -44,20 +44,18 @@
 #include <linux/prefetch.h>
 #include <linux/genalloc.h>
 #include <asm/cacheflush.h>
+#include <linux/io.h>
+#include <linux/delay.h>
 
 /* For 2-element tables related to cache-inhibited and cache-enabled mappings */
 #define DPAA_PORTAL_CE 0
 #define DPAA_PORTAL_CI 1
 
-#if (L1_CACHE_BYTES != 32) && (L1_CACHE_BYTES != 64)
-#error "Unsupported Cacheline Size"
-#endif
-
 static inline void dpaa_flush(void *p)
 {
 #ifdef CONFIG_PPC
        flush_dcache_range((unsigned long)p, (unsigned long)p+64);
-#elif defined(CONFIG_ARM32)
+#elif defined(CONFIG_ARM)
        __cpuc_flush_dcache_area(p, 64);
 #elif defined(CONFIG_ARM64)
        __flush_dcache_area(p, 64);
--- a/drivers/soc/fsl/qbman/qman.c
+++ b/drivers/soc/fsl/qbman/qman.c
@@ -41,6 +41,43 @@
 
 /* Portal register assists */
 
+#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
+/* Cache-inhibited register offsets */
+#define QM_REG_EQCR_PI_CINH    0x3000
+#define QM_REG_EQCR_CI_CINH    0x3040
+#define QM_REG_EQCR_ITR                0x3080
+#define QM_REG_DQRR_PI_CINH    0x3100
+#define QM_REG_DQRR_CI_CINH    0x3140
+#define QM_REG_DQRR_ITR                0x3180
+#define QM_REG_DQRR_DCAP       0x31C0
+#define QM_REG_DQRR_SDQCR      0x3200
+#define QM_REG_DQRR_VDQCR      0x3240
+#define QM_REG_DQRR_PDQCR      0x3280
+#define QM_REG_MR_PI_CINH      0x3300
+#define QM_REG_MR_CI_CINH      0x3340
+#define QM_REG_MR_ITR          0x3380
+#define QM_REG_CFG             0x3500
+#define QM_REG_ISR             0x3600
+#define QM_REG_IER             0x3640
+#define QM_REG_ISDR            0x3680
+#define QM_REG_IIR             0x36C0
+#define QM_REG_ITPR            0x3740
+
+/* Cache-enabled register offsets */
+#define QM_CL_EQCR             0x0000
+#define QM_CL_DQRR             0x1000
+#define QM_CL_MR               0x2000
+#define QM_CL_EQCR_PI_CENA     0x3000
+#define QM_CL_EQCR_CI_CENA     0x3040
+#define QM_CL_DQRR_PI_CENA     0x3100
+#define QM_CL_DQRR_CI_CENA     0x3140
+#define QM_CL_MR_PI_CENA       0x3300
+#define QM_CL_MR_CI_CENA       0x3340
+#define QM_CL_CR               0x3800
+#define QM_CL_RR0              0x3900
+#define QM_CL_RR1              0x3940
+
+#else
 /* Cache-inhibited register offsets */
 #define QM_REG_EQCR_PI_CINH    0x0000
 #define QM_REG_EQCR_CI_CINH    0x0004
@@ -75,6 +112,7 @@
 #define QM_CL_CR               0x3800
 #define QM_CL_RR0              0x3900
 #define QM_CL_RR1              0x3940
+#endif
 
 /*
  * BTW, the drivers (and h/w programming model) already obtain the required
@@ -909,12 +947,12 @@ static inline int qm_mc_result_timeout(s
 
 static inline void fq_set(struct qman_fq *fq, u32 mask)
 {
-       set_bits(mask, &fq->flags);
+       fq->flags |= mask;
 }
 
 static inline void fq_clear(struct qman_fq *fq, u32 mask)
 {
-       clear_bits(mask, &fq->flags);
+       fq->flags &= ~mask;
 }
 
 static inline int fq_isset(struct qman_fq *fq, u32 mask)
@@ -1566,7 +1604,7 @@ void qman_p_irqsource_add(struct qman_po
        unsigned long irqflags;
 
        local_irq_save(irqflags);
-       set_bits(bits & QM_PIRQ_VISIBLE, &p->irq_sources);
+       p->irq_sources |= bits & QM_PIRQ_VISIBLE;
        qm_out(&p->p, QM_REG_IER, p->irq_sources);
        local_irq_restore(irqflags);
 }
@@ -1589,7 +1627,7 @@ void qman_p_irqsource_remove(struct qman
         */
        local_irq_save(irqflags);
        bits &= QM_PIRQ_VISIBLE;
-       clear_bits(bits, &p->irq_sources);
+       p->irq_sources &= ~bits;
        qm_out(&p->p, QM_REG_IER, p->irq_sources);
        ier = qm_in(&p->p, QM_REG_IER);
        /*
--- a/drivers/soc/fsl/qbman/qman_ccsr.c
+++ b/drivers/soc/fsl/qbman/qman_ccsr.c
@@ -401,21 +401,42 @@ static int qm_init_pfdr(struct device *d
 }
 
 /*
- * Ideally we would use the DMA API to turn rmem->base into a DMA address
- * (especially if iommu translations ever get involved).  Unfortunately, the
- * DMA API currently does not allow mapping anything that is not backed with
- * a struct page.
+ * QMan needs two global memory areas initialized at boot time:
+ *  1) FQD: Frame Queue Descriptors used to manage frame queues
+ *  2) PFDR: Packed Frame Queue Descriptor Records used to store frames
+ * Both areas are reserved using the device tree reserved memory framework
+ * and the addresses and sizes are initialized when the QMan device is probed
  */
 static dma_addr_t fqd_a, pfdr_a;
 static size_t fqd_sz, pfdr_sz;
 
+#ifdef CONFIG_PPC
+/*
+ * Support for PPC Device Tree backward compatibility when compatiable
+ * string is set to fsl-qman-fqd and fsl-qman-pfdr
+ */
+static int zero_priv_mem(phys_addr_t addr, size_t sz)
+{
+       /* map as cacheable, non-guarded */
+       void __iomem *tmpp = ioremap_prot(addr, sz, 0);
+
+       if (!tmpp)
+               return -ENOMEM;
+
+       memset_io(tmpp, 0, sz);
+       flush_dcache_range((unsigned long)tmpp,
+                          (unsigned long)tmpp + sz);
+       iounmap(tmpp);
+
+       return 0;
+}
+
 static int qman_fqd(struct reserved_mem *rmem)
 {
        fqd_a = rmem->base;
        fqd_sz = rmem->size;
 
        WARN_ON(!(fqd_a && fqd_sz));
-
        return 0;
 }
 RESERVEDMEM_OF_DECLARE(qman_fqd, "fsl,qman-fqd", qman_fqd);
@@ -431,32 +452,13 @@ static int qman_pfdr(struct reserved_mem
 }
 RESERVEDMEM_OF_DECLARE(qman_pfdr, "fsl,qman-pfdr", qman_pfdr);
 
+#endif
+
 static unsigned int qm_get_fqid_maxcnt(void)
 {
        return fqd_sz / 64;
 }
 
-/*
- * Flush this memory range from data cache so that QMAN originated
- * transactions for this memory region could be marked non-coherent.
- */
-static int zero_priv_mem(struct device *dev, struct device_node *node,
-                        phys_addr_t addr, size_t sz)
-{
-       /* map as cacheable, non-guarded */
-       void __iomem *tmpp = ioremap_prot(addr, sz, 0);
-
-       if (!tmpp)
-               return -ENOMEM;
-
-       memset_io(tmpp, 0, sz);
-       flush_dcache_range((unsigned long)tmpp,
-                          (unsigned long)tmpp + sz);
-       iounmap(tmpp);
-
-       return 0;
-}
-
 static void log_edata_bits(struct device *dev, u32 bit_count)
 {
        u32 i, j, mask = 0xffffffff;
@@ -687,11 +689,12 @@ static int qman_resource_init(struct dev
 static int fsl_qman_probe(struct platform_device *pdev)
 {
        struct device *dev = &pdev->dev;
-       struct device_node *node = dev->of_node;
+       struct device_node *mem_node, *node = dev->of_node;
        struct resource *res;
        int ret, err_irq;
        u16 id;
        u8 major, minor;
+       u64 size;
 
        res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
        if (!res) {
@@ -717,6 +720,8 @@ static int fsl_qman_probe(struct platfor
                qman_ip_rev = QMAN_REV30;
        else if (major == 3 && minor == 1)
                qman_ip_rev = QMAN_REV31;
+       else if (major == 3 && minor == 2)
+               qman_ip_rev = QMAN_REV32;
        else {
                dev_err(dev, "Unknown QMan version\n");
                return -ENODEV;
@@ -727,10 +732,83 @@ static int fsl_qman_probe(struct platfor
                qm_channel_caam = QMAN_CHANNEL_CAAM_REV3;
        }
 
-       ret = zero_priv_mem(dev, node, fqd_a, fqd_sz);
-       WARN_ON(ret);
-       if (ret)
-               return -ENODEV;
+       if (fqd_a) {
+#ifdef CONFIG_PPC
+               /*
+                * For PPC backward DT compatibility
+                * FQD memory MUST be zero'd by software
+                */
+               zero_priv_mem(fqd_a, fqd_sz);
+#else
+               WARN(1, "Unexpected archiceture using non shared-dma-mem reservations");
+#endif
+       } else {
+               /*
+                * Order of memory regions is assumed as FQD followed by PFDR
+                * in order to ensure allocations from the correct regions the
+                * driver initializes then allocates each piece in order
+                */
+               ret = of_reserved_mem_device_init_by_idx(dev, dev->of_node, 0);
+               if (ret) {
+                       dev_err(dev, "of_reserved_mem_device_init_by_idx(0) failed 0x%x\n",
+                               ret);
+                       return -ENODEV;
+               }
+               mem_node = of_parse_phandle(dev->of_node, "memory-region", 0);
+               if (mem_node) {
+                       ret = of_property_read_u64(mem_node, "size", &size);
+                       if (ret) {
+                               dev_err(dev, "FQD: of_address_to_resource fails 0x%x\n",
+                                       ret);
+                               return -ENODEV;
+                       }
+                       fqd_sz = size;
+               } else {
+                       dev_err(dev, "No memory-region found for FQD\n");
+                       return -ENODEV;
+               }
+               if (!dma_zalloc_coherent(dev, fqd_sz, &fqd_a, 0)) {
+                       dev_err(dev, "Alloc FQD memory failed\n");
+                       return -ENODEV;
+               }
+
+               /*
+                * Disassociate the FQD reserved memory area from the device
+                * because a device can only have one DMA memory area. This
+                * should be fine since the memory is allocated and initialized
+                * and only ever accessed by the QMan device from now on
+                */
+               of_reserved_mem_device_release(dev);
+       }
+       dev_dbg(dev, "Allocated FQD 0x%llx 0x%zx\n", fqd_a, fqd_sz);
+
+       if (!pfdr_a) {
+               /* Setup PFDR memory */
+               ret = of_reserved_mem_device_init_by_idx(dev, dev->of_node, 1);
+               if (ret) {
+                       dev_err(dev, "of_reserved_mem_device_init(1) failed 0x%x\n",
+                       ret);
+                       return -ENODEV;
+               }
+               mem_node = of_parse_phandle(dev->of_node, "memory-region", 1);
+               if (mem_node) {
+                       ret = of_property_read_u64(mem_node, "size", &size);
+                       if (ret) {
+                               dev_err(dev, "PFDR: of_address_to_resource fails 0x%x\n",
+                                       ret);
+                               return -ENODEV;
+                       }
+                       pfdr_sz = size;
+               } else {
+                       dev_err(dev, "No memory-region found for PFDR\n");
+                       return -ENODEV;
+               }
+               if (!dma_zalloc_coherent(dev, pfdr_sz, &pfdr_a, 0)) {
+                       dev_err(dev, "Alloc PFDR Failed size 0x%zx\n", pfdr_sz);
+                       return -ENODEV;
+               }
+       }
+       dev_info(dev, "Allocated PFDR 0x%llx 0x%zx\n", pfdr_a, pfdr_sz);
 
        ret = qman_init_ccsr(dev);
        if (ret) {
--- a/drivers/soc/fsl/qbman/qman_portal.c
+++ b/drivers/soc/fsl/qbman/qman_portal.c
@@ -262,7 +262,14 @@ static int qman_portal_probe(struct plat
        }
        pcfg->irq = irq;
 
-       va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]), 0);
+#ifdef CONFIG_PPC
+       /* PPC requires a cacheable/non-coherent mapping of the portal */
+       va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]),
+                         (pgprot_val(PAGE_KERNEL) & ~_PAGE_COHERENT));
+#else
+       /* For ARM we can use write combine mapping. */
+       va = ioremap_wc(addr_phys[0]->start, resource_size(addr_phys[0]));
+#endif
        if (!va) {
                dev_err(dev, "ioremap::CE failed\n");
                goto err_ioremap1;
@@ -270,8 +277,7 @@ static int qman_portal_probe(struct plat
 
        pcfg->addr_virt[DPAA_PORTAL_CE] = va;
 
-       va = ioremap_prot(addr_phys[1]->start, resource_size(addr_phys[1]),
-                         _PAGE_GUARDED | _PAGE_NO_CACHE);
+       va = ioremap(addr_phys[1]->start, resource_size(addr_phys[1]));
        if (!va) {
                dev_err(dev, "ioremap::CI failed\n");
                goto err_ioremap2;
--- a/drivers/soc/fsl/qbman/qman_priv.h
+++ b/drivers/soc/fsl/qbman/qman_priv.h
@@ -28,13 +28,13 @@
  * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
  */
 
-#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-
 #include "dpaa_sys.h"
 
 #include <soc/fsl/qman.h>
 #include <linux/dma-mapping.h>
 #include <linux/iommu.h>
+#include <linux/dma-contiguous.h>
+#include <linux/of_address.h>
 
 #if defined(CONFIG_FSL_PAMU)
 #include <asm/fsl_pamu_stash.h>
@@ -187,6 +187,7 @@ struct qm_portal_config {
 #define QMAN_REV20 0x0200
 #define QMAN_REV30 0x0300
 #define QMAN_REV31 0x0301
+#define QMAN_REV32 0x0302
 extern u16 qman_ip_rev; /* 0 if uninitialised, otherwise QMAN_REVx */
 
 #define QM_FQID_RANGE_START 1 /* FQID 0 reserved for internal use */
--- a/drivers/soc/fsl/qbman/qman_test.h
+++ b/drivers/soc/fsl/qbman/qman_test.h
@@ -30,7 +30,5 @@
 
 #include "qman_priv.h"
 
-#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-
 int qman_test_stash(void);
 int qman_test_api(void);