diff options
Diffstat (limited to 'target/linux/layerscape')
57 files changed, 19862 insertions, 1 deletions
diff --git a/target/linux/layerscape/64b/profiles/00-default.mk b/target/linux/layerscape/64b/profiles/00-default.mk index a2368af..68b18ba 100644 --- a/target/linux/layerscape/64b/profiles/00-default.mk +++ b/target/linux/layerscape/64b/profiles/00-default.mk @@ -9,6 +9,8 @@ endef DEFAULT_PACKAGES+= \ rcw-layerscape-ls1043ardb uboot-layerscape-$(SUBTARGET)-ls1043ardb \ - fman-layerscape-ls1043ardb + fman-layerscape-ls1043ardb \ + rcw-layerscape-ls1012ardb uboot-layerscape-$(SUBTARGET)-ls1012ardb \ + kmod-ppfe ppfe-ls1012ardb $(eval $(call Profile,Default)) diff --git a/target/linux/layerscape/base-files/etc/rc.local b/target/linux/layerscape/base-files/etc/rc.local new file mode 100644 index 0000000..297c3a8 --- /dev/null +++ b/target/linux/layerscape/base-files/etc/rc.local @@ -0,0 +1,4 @@ +# Put your custom commands here that should be executed once +# the system init finished. By default this file does nothing. +modprobe pfe.ko +exit 0 diff --git a/target/linux/layerscape/config-4.4 b/target/linux/layerscape/config-4.4 index daf2013..88da774 100644 --- a/target/linux/layerscape/config-4.4 +++ b/target/linux/layerscape/config-4.4 @@ -294,3 +294,5 @@ CONFIG_USB=y CONFIG_VITESSE_PHY=y CONFIG_XPS=y CONFIG_ZLIB_INFLATE=y +CONFIG_MTD_SPI_NOR=y +CONFIG_SPI_FSL_QUADSPI=y diff --git a/target/linux/layerscape/image/Makefile b/target/linux/layerscape/image/Makefile index 9153f00..7d80398 100644 --- a/target/linux/layerscape/image/Makefile +++ b/target/linux/layerscape/image/Makefile @@ -25,6 +25,11 @@ define Build/append-ls-dtb dd if=$(DTS_DIR)/$(1).dtb >> $@ endef +define Build/append-ls-rootfs-ext4 + $(STAGING_DIR_HOST)/bin/make_ext4fs -l $(word 2,$(1)) -b 4096 -i 6000 -m 0 -J $(KDIR)/$(word 1,$(1))-$(word 2,$(1)).root.ext4 $(TARGET_DIR) + dd if=$(KDIR)/$(word 1,$(1))-$(word 2,$(1)).root.ext4 >> $@ +endef + define Device/Default PROFILES = Default FILESYSTEMS := squashfs @@ -57,4 +62,23 @@ endif endef TARGET_DEVICES += ls1043ardb +define Device/ls1012ardb + DEVICE_TITLE := ls1012ardb-$(SUBTARGET) + DEVICE_PACKAGES += rcw-layerscape-ls1012ardb uboot-layerscape-$(SUBTARGET)-ls1012ardb kmod-ppfe ppfe-ls1012ardb +ifeq ($(SUBTARGET),64b) + DEVICE_DTS = freescale/fsl-ls1012a-rdb +endif +ifeq ($(SUBTARGET),32b) + DEVICE_DTS = ../../../arm64/boot/dts/freescale/fsl-ls1012a-rdb +endif + IMAGE/firmware.bin = append-ls-rcw $(1) | pad-to 1M | append-ls-uboot $(1) | pad-to 3M | \ + append-ls-dtb $$(DEVICE_DTS) | pad-to 4M | append-kernel | pad-to 9M | \ + append-rootfs | pad-to 32M | check-size 33554433 + IMAGES += firmware.ext4.bin + IMAGE/firmware.ext4.bin = append-ls-rcw $(1) | pad-to 1M | append-ls-uboot $(1) | pad-to 3M | \ + append-ls-dtb $$(DEVICE_DTS) | pad-to 4M | append-kernel | pad-to 9M | \ + append-ls-rootfs-ext4 $(1) 23M | check-size 33554433 +endef +TARGET_DEVICES += ls1012ardb + $(eval $(call BuildImage)) diff --git a/target/linux/layerscape/modules.mk b/target/linux/layerscape/modules.mk new file mode 100644 index 0000000..0b34f96 --- /dev/null +++ b/target/linux/layerscape/modules.mk @@ -0,0 +1,20 @@ +# +# Copyright (C) Jiang Yutang <jiangyutang1978@gmail.com> +# +# This is free software, licensed under the GNU General Public License v2. +# See /LICENSE for more information. +# + +define KernelPackage/ppfe + SUBMENU:=$(NETWORK_DEVICES_MENU) + TITLE:=Freescale PPFE Driver support + KCONFIG:=CONFIG_FSL_PPFE + FILES:=$(LINUX_DIR)/drivers/staging/fsl_ppfe/pfe.ko + AUTOLOAD:=$(call AutoLoad,35,ppfe) +endef + +define KernelPackage/ppfe/description + Kernel modules for Freescale PPFE Driver support. +endef + +$(eval $(call KernelPackage,ppfe)) diff --git a/target/linux/layerscape/patches-4.4/1074-mtd-nand-spi-nor-assign-MTD-of_node.patch b/target/linux/layerscape/patches-4.4/1074-mtd-nand-spi-nor-assign-MTD-of_node.patch new file mode 100644 index 0000000..4130e84 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1074-mtd-nand-spi-nor-assign-MTD-of_node.patch @@ -0,0 +1,31 @@ +From f560fdb9d71aaf3adc54341a1650577c78495df9 Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Fri, 30 Oct 2015 20:33:22 -0700 +Subject: [PATCH 074/113] mtd: {nand,spi-nor}: assign MTD of_node + +We should pass along our flash DT node to the MTD layer, so it can set +up ofpart for us. + +cherry-pick{ +remove the code: +drivers/mtd/nand/nand_base.c | 3 + +commit:3e63b26bdd4069c3df2cd7ce7217a21d06801b41 +} + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Reviewed-by: Boris Brezillon <boris.brezillon@free-electrons.com> +Signed-off-by: Yuan Yao <yao.yuan@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 1 + + 1 file changed, 1 insertion(+) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -1217,6 +1217,7 @@ int spi_nor_scan(struct spi_nor *nor, co + mtd->flags |= MTD_NO_ERASE; + + mtd->dev.parent = dev; ++ mtd_set_of_node(mtd, np); + nor->page_size = info->page_size; + mtd->writebufsize = nor->page_size; + diff --git a/target/linux/layerscape/patches-4.4/1075-mtd-spi-nor-convert-to-spi_nor_-get-set-_flash_node.patch b/target/linux/layerscape/patches-4.4/1075-mtd-spi-nor-convert-to-spi_nor_-get-set-_flash_node.patch new file mode 100644 index 0000000..182efac --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1075-mtd-spi-nor-convert-to-spi_nor_-get-set-_flash_node.patch @@ -0,0 +1,80 @@ +From f906ec330da9aa83de5382653436be36273c63d3 Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Fri, 30 Oct 2015 20:33:24 -0700 +Subject: [PATCH 075/113] mtd: spi-nor: convert to spi_nor_{get, + set}_flash_node() + +Used semantic patch with 'make coccicheck MODE=patch COCCI=script.cocci': + +---8<---- +virtual patch + +@@ +struct spi_nor b; +struct spi_nor *c; +expression d; +@@ +( +-(b).flash_node = (d) ++spi_nor_set_flash_node(&b, d) +| +-(c)->flash_node = (d) ++spi_nor_set_flash_node(c, d) +) +---8<---- + +And a manual conversion for the one use of spi_nor_get_flash_node(). + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Reviewed-by: Boris Brezillon <boris.brezillon@free-electrons.com> +--- + drivers/mtd/devices/m25p80.c | 2 +- + drivers/mtd/spi-nor/fsl-quadspi.c | 2 +- + drivers/mtd/spi-nor/nxp-spifi.c | 2 +- + drivers/mtd/spi-nor/spi-nor.c | 2 +- + 4 files changed, 4 insertions(+), 4 deletions(-) + +--- a/drivers/mtd/devices/m25p80.c ++++ b/drivers/mtd/devices/m25p80.c +@@ -221,7 +221,7 @@ static int m25p_probe(struct spi_device + nor->read_reg = m25p80_read_reg; + + nor->dev = &spi->dev; +- nor->flash_node = spi->dev.of_node; ++ spi_nor_set_flash_node(nor, spi->dev.of_node); + nor->priv = flash; + + spi_set_drvdata(spi, flash); +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -1013,7 +1013,7 @@ static int fsl_qspi_probe(struct platfor + mtd = &nor->mtd; + + nor->dev = dev; +- nor->flash_node = np; ++ spi_nor_set_flash_node(nor, np); + nor->priv = q; + + /* fill the hooks */ +--- a/drivers/mtd/spi-nor/nxp-spifi.c ++++ b/drivers/mtd/spi-nor/nxp-spifi.c +@@ -330,7 +330,7 @@ static int nxp_spifi_setup_flash(struct + writel(ctrl, spifi->io_base + SPIFI_CTRL); + + spifi->nor.dev = spifi->dev; +- spifi->nor.flash_node = np; ++ spi_nor_set_flash_node(&spifi->nor, np); + spifi->nor.priv = spifi; + spifi->nor.read = nxp_spifi_read; + spifi->nor.write = nxp_spifi_write; +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -1111,7 +1111,7 @@ int spi_nor_scan(struct spi_nor *nor, co + const struct flash_info *info = NULL; + struct device *dev = nor->dev; + struct mtd_info *mtd = &nor->mtd; +- struct device_node *np = nor->flash_node; ++ struct device_node *np = spi_nor_get_flash_node(nor); + int ret; + int i; + diff --git a/target/linux/layerscape/patches-4.4/1076-mtd-spi-nor-drop-unnecessary-partition-parser-data.patch b/target/linux/layerscape/patches-4.4/1076-mtd-spi-nor-drop-unnecessary-partition-parser-data.patch new file mode 100644 index 0000000..0960f53 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1076-mtd-spi-nor-drop-unnecessary-partition-parser-data.patch @@ -0,0 +1,83 @@ +From e36da6d0a0841ea3a75d5189057bd020d737e71a Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Fri, 30 Oct 2015 20:33:26 -0700 +Subject: [PATCH 076/113] mtd: spi-nor: drop unnecessary partition parser data + +Now that the SPI-NOR/MTD framework pass the 'flash_node' through to the +partition parsing code, we don't have to do it ourselves. + +Also convert to mtd_device_register(), since we don't need the 2nd and +3rd parameters anymore. + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Reviewed-by: Boris Brezillon <boris.brezillon@free-electrons.com> +--- + drivers/mtd/devices/m25p80.c | 8 ++------ + drivers/mtd/spi-nor/fsl-quadspi.c | 4 +--- + drivers/mtd/spi-nor/nxp-spifi.c | 4 +--- + 3 files changed, 4 insertions(+), 12 deletions(-) + +--- a/drivers/mtd/devices/m25p80.c ++++ b/drivers/mtd/devices/m25p80.c +@@ -197,7 +197,6 @@ static int m25p80_erase(struct spi_nor * + */ + static int m25p_probe(struct spi_device *spi) + { +- struct mtd_part_parser_data ppdata; + struct flash_platform_data *data; + struct m25p *flash; + struct spi_nor *nor; +@@ -249,11 +248,8 @@ static int m25p_probe(struct spi_device + if (ret) + return ret; + +- ppdata.of_node = spi->dev.of_node; +- +- return mtd_device_parse_register(&nor->mtd, NULL, &ppdata, +- data ? data->parts : NULL, +- data ? data->nr_parts : 0); ++ return mtd_device_register(&nor->mtd, data ? data->parts : NULL, ++ data ? data->nr_parts : 0); + } + + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -927,7 +927,6 @@ static void fsl_qspi_unprep(struct spi_n + static int fsl_qspi_probe(struct platform_device *pdev) + { + struct device_node *np = pdev->dev.of_node; +- struct mtd_part_parser_data ppdata; + struct device *dev = &pdev->dev; + struct fsl_qspi *q; + struct resource *res; +@@ -1038,8 +1037,7 @@ static int fsl_qspi_probe(struct platfor + if (ret) + goto mutex_failed; + +- ppdata.of_node = np; +- ret = mtd_device_parse_register(mtd, NULL, &ppdata, NULL, 0); ++ ret = mtd_device_register(mtd, NULL, 0); + if (ret) + goto mutex_failed; + +--- a/drivers/mtd/spi-nor/nxp-spifi.c ++++ b/drivers/mtd/spi-nor/nxp-spifi.c +@@ -271,7 +271,6 @@ static void nxp_spifi_dummy_id_read(stru + static int nxp_spifi_setup_flash(struct nxp_spifi *spifi, + struct device_node *np) + { +- struct mtd_part_parser_data ppdata; + enum read_mode flash_read; + u32 ctrl, property; + u16 mode = 0; +@@ -361,8 +360,7 @@ static int nxp_spifi_setup_flash(struct + return ret; + } + +- ppdata.of_node = np; +- ret = mtd_device_parse_register(&spifi->nor.mtd, NULL, &ppdata, NULL, 0); ++ ret = mtd_device_register(&spifi->nor.mtd, NULL, 0); + if (ret) { + dev_err(spifi->dev, "mtd device parse failed\n"); + return ret; diff --git a/target/linux/layerscape/patches-4.4/1077-mtd-add-get-set-of_node-flash_node-helpers.patch b/target/linux/layerscape/patches-4.4/1077-mtd-add-get-set-of_node-flash_node-helpers.patch new file mode 100644 index 0000000..eeba1e9 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1077-mtd-add-get-set-of_node-flash_node-helpers.patch @@ -0,0 +1,62 @@ +From a2f87e7df641b482e217f5b0efbaf41f6b8a0cf6 Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Fri, 30 Oct 2015 20:33:20 -0700 +Subject: [PATCH 077/113] mtd: add get/set of_node/flash_node helpers + +We are going to begin using the mtd->dev.of_node field for MTD device +nodes, so let's add helpers for it. Also, we'll be making some +conversions on spi_nor (and nand_chip eventually) too, so get that ready +with their own helpers. + +commit:28b8b26b308e656edfa9467867d5f79212da2ec3 +delete the include/linux/mtd/nand.h +just upgrade the code about spi. + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Reviewed-by: Boris Brezillon <boris.brezillon@free-electrons.com> +Signed-off-by: Yuan Yao <yao.yuan@nxp.com> +--- + include/linux/mtd/mtd.h | 11 +++++++++++ + include/linux/mtd/spi-nor.h | 11 +++++++++++ + 2 files changed, 22 insertions(+) + +--- a/include/linux/mtd/mtd.h ++++ b/include/linux/mtd/mtd.h +@@ -258,6 +258,17 @@ struct mtd_info { + int usecount; + }; + ++static inline void mtd_set_of_node(struct mtd_info *mtd, ++ struct device_node *np) ++{ ++ mtd->dev.of_node = np; ++} ++ ++static inline struct device_node *mtd_get_of_node(struct mtd_info *mtd) ++{ ++ return mtd->dev.of_node; ++} ++ + int mtd_erase(struct mtd_info *mtd, struct erase_info *instr); + int mtd_point(struct mtd_info *mtd, loff_t from, size_t len, size_t *retlen, + void **virt, resource_size_t *phys); +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -184,6 +184,17 @@ struct spi_nor { + void *priv; + }; + ++static inline void spi_nor_set_flash_node(struct spi_nor *nor, ++ struct device_node *np) ++{ ++ nor->flash_node = np; ++} ++ ++static inline struct device_node *spi_nor_get_flash_node(struct spi_nor *nor) ++{ ++ return nor->flash_node; ++} ++ + /** + * spi_nor_scan() - scan the SPI NOR + * @nor: the spi_nor structure diff --git a/target/linux/layerscape/patches-4.4/1078-mtd-spi-nor-drop-flash_node-field.patch b/target/linux/layerscape/patches-4.4/1078-mtd-spi-nor-drop-flash_node-field.patch new file mode 100644 index 0000000..baaf9e5 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1078-mtd-spi-nor-drop-flash_node-field.patch @@ -0,0 +1,57 @@ +From df36b4601bc9f84684249a26eb39b818d6785fb8 Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Fri, 30 Oct 2015 20:33:27 -0700 +Subject: [PATCH 078/113] mtd: spi-nor: drop flash_node field + +We can just alias to the MTD of_node. + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Reviewed-by: Boris Brezillon <boris.brezillon@free-electrons.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 1 - + include/linux/mtd/spi-nor.h | 6 ++---- + 2 files changed, 2 insertions(+), 5 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -1217,7 +1217,6 @@ int spi_nor_scan(struct spi_nor *nor, co + mtd->flags |= MTD_NO_ERASE; + + mtd->dev.parent = dev; +- mtd_set_of_node(mtd, np); + nor->page_size = info->page_size; + mtd->writebufsize = nor->page_size; + +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -123,7 +123,6 @@ enum spi_nor_option_flags { + * @mtd: point to a mtd_info structure + * @lock: the lock for the read/write/erase/lock/unlock operations + * @dev: point to a spi device, or a spi nor controller device. +- * @flash_node: point to a device node describing this flash instance. + * @page_size: the page size of the SPI NOR + * @addr_width: number of address bytes + * @erase_opcode: the opcode for erasing a sector +@@ -154,7 +153,6 @@ struct spi_nor { + struct mtd_info mtd; + struct mutex lock; + struct device *dev; +- struct device_node *flash_node; + u32 page_size; + u8 addr_width; + u8 erase_opcode; +@@ -187,12 +185,12 @@ struct spi_nor { + static inline void spi_nor_set_flash_node(struct spi_nor *nor, + struct device_node *np) + { +- nor->flash_node = np; ++ mtd_set_of_node(&nor->mtd, np); + } + + static inline struct device_node *spi_nor_get_flash_node(struct spi_nor *nor) + { +- return nor->flash_node; ++ return mtd_get_of_node(&nor->mtd); + } + + /** diff --git a/target/linux/layerscape/patches-4.4/1079-mtd-spi-nor-remove-unnecessary-leading-space-from-db.patch b/target/linux/layerscape/patches-4.4/1079-mtd-spi-nor-remove-unnecessary-leading-space-from-db.patch new file mode 100644 index 0000000..159e7fd --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1079-mtd-spi-nor-remove-unnecessary-leading-space-from-db.patch @@ -0,0 +1,27 @@ +From 3ea419cf269832f5743d9b5ad75ece5178b02b09 Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Fri, 30 Oct 2015 12:56:22 -0700 +Subject: [PATCH 079/113] mtd: spi-nor: remove unnecessary leading space from + dbg print + +As Cyrille noted [1], this line is wrong. + +[1] http://lists.infradead.org/pipermail/linux-mtd/2015-September/061725.html + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Cc: Cyrille Pitchen <cyrille.pitchen@atmel.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -860,7 +860,7 @@ static const struct flash_info *spi_nor_ + + tmp = nor->read_reg(nor, SPINOR_OP_RDID, id, SPI_NOR_MAX_ID_LEN); + if (tmp < 0) { +- dev_dbg(nor->dev, " error %d reading JEDEC ID\n", tmp); ++ dev_dbg(nor->dev, "error %d reading JEDEC ID\n", tmp); + return ERR_PTR(tmp); + } + diff --git a/target/linux/layerscape/patches-4.4/1080-mtd-fsl-quadspi-possible-NULL-dereference.patch b/target/linux/layerscape/patches-4.4/1080-mtd-fsl-quadspi-possible-NULL-dereference.patch new file mode 100644 index 0000000..76898ab --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1080-mtd-fsl-quadspi-possible-NULL-dereference.patch @@ -0,0 +1,50 @@ +From bd02decd1ad7cc883ce388e769a34a3c402b90c4 Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Mon, 16 Nov 2015 10:45:30 -0800 +Subject: [PATCH 080/113] mtd: fsl-quadspi: possible NULL dereference + +It is theoretically possible to probe this driver without a matching +device tree, so let's guard against this. + +Also, use the of_device_get_match_data() helper to make this a bit +simpler. + +Coverity complained about this one. + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Acked-by: Han xu <han.xu@freescale.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 8 ++++---- + 1 file changed, 4 insertions(+), 4 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -269,7 +269,7 @@ struct fsl_qspi { + struct clk *clk, *clk_en; + struct device *dev; + struct completion c; +- struct fsl_qspi_devtype_data *devtype_data; ++ const struct fsl_qspi_devtype_data *devtype_data; + u32 nor_size; + u32 nor_num; + u32 clk_rate; +@@ -933,8 +933,6 @@ static int fsl_qspi_probe(struct platfor + struct spi_nor *nor; + struct mtd_info *mtd; + int ret, i = 0; +- const struct of_device_id *of_id = +- of_match_device(fsl_qspi_dt_ids, &pdev->dev); + + q = devm_kzalloc(dev, sizeof(*q), GFP_KERNEL); + if (!q) +@@ -945,7 +943,9 @@ static int fsl_qspi_probe(struct platfor + return -ENODEV; + + q->dev = dev; +- q->devtype_data = (struct fsl_qspi_devtype_data *)of_id->data; ++ q->devtype_data = of_device_get_match_data(dev); ++ if (!q->devtype_data) ++ return -ENODEV; + platform_set_drvdata(pdev, q); + + /* find the resources */ diff --git a/target/linux/layerscape/patches-4.4/1081-mtd-spi-nor-provide-default-erase_sector-implementat.patch b/target/linux/layerscape/patches-4.4/1081-mtd-spi-nor-provide-default-erase_sector-implementat.patch new file mode 100644 index 0000000..8becdd2 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1081-mtd-spi-nor-provide-default-erase_sector-implementat.patch @@ -0,0 +1,105 @@ +From 56bd0e13d8bc3b4486251b10ac9d2ba7434c21ee Mon Sep 17 00:00:00 2001 +From: Brian Norris <computersforpeace@gmail.com> +Date: Tue, 10 Nov 2015 12:15:27 -0800 +Subject: [PATCH 081/113] mtd: spi-nor: provide default erase_sector + implementation + +Some spi-nor drivers perform sector erase by duplicating their +write_reg() command. Let's not require that the driver fill this out, +and provide a default instead. + +Tested on m25p80.c and Medatek's MT8173 SPI NOR flash driver. + +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 37 +++++++++++++++++++++++++++++++++---- + include/linux/mtd/spi-nor.h | 3 ++- + 2 files changed, 35 insertions(+), 5 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -38,6 +38,7 @@ + #define CHIP_ERASE_2MB_READY_WAIT_JIFFIES (40UL * HZ) + + #define SPI_NOR_MAX_ID_LEN 6 ++#define SPI_NOR_MAX_ADDR_WIDTH 4 + + struct flash_info { + char *name; +@@ -313,6 +314,29 @@ static void spi_nor_unlock_and_unprep(st + } + + /* ++ * Initiate the erasure of a single sector ++ */ ++static int spi_nor_erase_sector(struct spi_nor *nor, u32 addr) ++{ ++ u8 buf[SPI_NOR_MAX_ADDR_WIDTH]; ++ int i; ++ ++ if (nor->erase) ++ return nor->erase(nor, addr); ++ ++ /* ++ * Default implementation, if driver doesn't have a specialized HW ++ * control ++ */ ++ for (i = nor->addr_width - 1; i >= 0; i--) { ++ buf[i] = addr & 0xff; ++ addr >>= 8; ++ } ++ ++ return nor->write_reg(nor, nor->erase_opcode, buf, nor->addr_width); ++} ++ ++/* + * Erase an address range on the nor chip. The address range may extend + * one or more erase sectors. Return an error is there is a problem erasing. + */ +@@ -371,10 +395,9 @@ static int spi_nor_erase(struct mtd_info + while (len) { + write_enable(nor); + +- if (nor->erase(nor, addr)) { +- ret = -EIO; ++ ret = spi_nor_erase_sector(nor, addr); ++ if (ret) + goto erase_err; +- } + + addr += mtd->erasesize; + len -= mtd->erasesize; +@@ -1098,7 +1121,7 @@ static int set_quad_mode(struct spi_nor + static int spi_nor_check(struct spi_nor *nor) + { + if (!nor->dev || !nor->read || !nor->write || +- !nor->read_reg || !nor->write_reg || !nor->erase) { ++ !nor->read_reg || !nor->write_reg) { + pr_err("spi-nor: please fill all the necessary fields!\n"); + return -EINVAL; + } +@@ -1299,6 +1322,12 @@ int spi_nor_scan(struct spi_nor *nor, co + nor->addr_width = 3; + } + ++ if (nor->addr_width > SPI_NOR_MAX_ADDR_WIDTH) { ++ dev_err(dev, "address width is too large: %u\n", ++ nor->addr_width); ++ return -EINVAL; ++ } ++ + nor->read_dummy = spi_nor_read_dummy_cycles(nor); + + dev_info(dev, "%s (%lld Kbytes)\n", info->name, +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -142,7 +142,8 @@ enum spi_nor_option_flags { + * @read: [DRIVER-SPECIFIC] read data from the SPI NOR + * @write: [DRIVER-SPECIFIC] write data to the SPI NOR + * @erase: [DRIVER-SPECIFIC] erase a sector of the SPI NOR +- * at the offset @offs ++ * at the offset @offs; if not provided by the driver, ++ * spi-nor will send the erase opcode via write_reg() + * @flash_lock: [FLASH-SPECIFIC] lock a region of the SPI NOR + * @flash_unlock: [FLASH-SPECIFIC] unlock a region of the SPI NOR + * @flash_is_locked: [FLASH-SPECIFIC] check if a region of the SPI NOR is diff --git a/target/linux/layerscape/patches-4.4/1083-mtd-spi-nor-Fix-error-message-with-unrecognized-JEDE.patch b/target/linux/layerscape/patches-4.4/1083-mtd-spi-nor-Fix-error-message-with-unrecognized-JEDE.patch new file mode 100644 index 0000000..5b98ff4 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1083-mtd-spi-nor-Fix-error-message-with-unrecognized-JEDE.patch @@ -0,0 +1,31 @@ +From 30e609daed95664824e95344e85c7eaedd1bfcf3 Mon Sep 17 00:00:00 2001 +From: Ricardo Ribalda <ricardo.ribalda@gmail.com> +Date: Mon, 30 Nov 2015 20:41:17 +0100 +Subject: [PATCH 083/113] mtd: spi-nor: Fix error message with unrecognized + JEDEC + +The error message was: + +m25p80 spi32766.0: unrecognized JEDEC id bytes: 00, 0, 0 + +The new error message: + +m25p80 spi32766.0: unrecognized JEDEC id bytes: 00, 00, 00 + +Signed-off-by: Ricardo Ribalda Delgado <ricardo.ribalda@gmail.com> +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -894,7 +894,7 @@ static const struct flash_info *spi_nor_ + return &spi_nor_ids[tmp]; + } + } +- dev_err(nor->dev, "unrecognized JEDEC id bytes: %02x, %2x, %2x\n", ++ dev_err(nor->dev, "unrecognized JEDEC id bytes: %02x, %02x, %02x\n", + id[0], id[1], id[2]); + return ERR_PTR(-ENODEV); + } diff --git a/target/linux/layerscape/patches-4.4/1084-mtd-spi-nor-fix-error-handling-in-spi_nor_erase.patch b/target/linux/layerscape/patches-4.4/1084-mtd-spi-nor-fix-error-handling-in-spi_nor_erase.patch new file mode 100644 index 0000000..3930028 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1084-mtd-spi-nor-fix-error-handling-in-spi_nor_erase.patch @@ -0,0 +1,39 @@ +From 9e473594776da97245049019f1d1e9608ff1214a Mon Sep 17 00:00:00 2001 +From: Heiner Kallweit <hkallweit1@gmail.com> +Date: Tue, 17 Nov 2015 20:18:54 +0100 +Subject: [PATCH 084/113] mtd: spi-nor: fix error handling in spi_nor_erase + +The documenting comment of mtd_erase in mtdcore.c states: +Device drivers are supposed to call instr->callback() whenever +the operation completes, even if it completes with a failure. + +Currently the callback isn't called in case of failure. Fix this. + +Signed-off-by: Heiner Kallweit <hkallweit1@gmail.com> +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 8 ++------ + 1 file changed, 2 insertions(+), 6 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -410,17 +410,13 @@ static int spi_nor_erase(struct mtd_info + + write_disable(nor); + ++erase_err: + spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_ERASE); + +- instr->state = MTD_ERASE_DONE; ++ instr->state = ret ? MTD_ERASE_FAILED : MTD_ERASE_DONE; + mtd_erase_callback(instr); + + return ret; +- +-erase_err: +- spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_ERASE); +- instr->state = MTD_ERASE_FAILED; +- return ret; + } + + static void stm_get_locked_range(struct spi_nor *nor, u8 sr, loff_t *ofs, diff --git a/target/linux/layerscape/patches-4.4/1085-mtd-spi-nor-Check-the-return-value-from-read_sr.patch b/target/linux/layerscape/patches-4.4/1085-mtd-spi-nor-Check-the-return-value-from-read_sr.patch new file mode 100644 index 0000000..6f39b6b --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1085-mtd-spi-nor-Check-the-return-value-from-read_sr.patch @@ -0,0 +1,58 @@ +From d05c68e35f42a46b352d2a4bdaef9954c946e20a Mon Sep 17 00:00:00 2001 +From: Fabio Estevam <fabio.estevam@freescale.com> +Date: Fri, 20 Nov 2015 16:26:11 -0200 +Subject: [PATCH 085/113] mtd: spi-nor: Check the return value from read_sr() + +[context adjustment] + +We should better check the return value from read_sr() and +propagate it in the case of error. + +Signed-off-by: Fabio Estevam <fabio.estevam@freescale.com> +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 10 ++++++++-- + 1 file changed, 8 insertions(+), 2 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -478,11 +478,13 @@ static int stm_is_locked_sr(struct spi_n + static int stm_lock(struct spi_nor *nor, loff_t ofs, uint64_t len) + { + struct mtd_info *mtd = &nor->mtd; +- u8 status_old, status_new; ++ int status_old, status_new; + u8 mask = SR_BP2 | SR_BP1 | SR_BP0; + u8 shift = ffs(mask) - 1, pow, val; + + status_old = read_sr(nor); ++ if (status_old < 0) ++ return status_old; + + /* SPI NOR always locks to the end */ + if (ofs + len != mtd->size) { +@@ -528,11 +530,13 @@ static int stm_lock(struct spi_nor *nor, + static int stm_unlock(struct spi_nor *nor, loff_t ofs, uint64_t len) + { + struct mtd_info *mtd = &nor->mtd; +- uint8_t status_old, status_new; ++ int status_old, status_new; + u8 mask = SR_BP2 | SR_BP1 | SR_BP0; + u8 shift = ffs(mask) - 1, pow, val; + + status_old = read_sr(nor); ++ if (status_old < 0) ++ return status_old; + + /* Cannot unlock; would unlock larger region than requested */ + if (stm_is_locked_sr(nor, ofs - mtd->erasesize, mtd->erasesize, +@@ -1036,6 +1040,8 @@ static int macronix_quad_enable(struct s + int ret, val; + + val = read_sr(nor); ++ if (val < 0) ++ return val; + write_enable(nor); + + write_sr(nor, val | SR_QUAD_EN_MX); diff --git a/target/linux/layerscape/patches-4.4/1086-mtd-spi-nor-wait-until-lock-unlock-operations-are-re.patch b/target/linux/layerscape/patches-4.4/1086-mtd-spi-nor-wait-until-lock-unlock-operations-are-re.patch new file mode 100644 index 0000000..7bcfd51 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1086-mtd-spi-nor-wait-until-lock-unlock-operations-are-re.patch @@ -0,0 +1,66 @@ +From 3a06c61b48fbc23046928275e37a693e1055ae74 Mon Sep 17 00:00:00 2001 +From: =?UTF-8?q?Ezequiel=20Garc=C3=ADa?= <ezequiel@vanguardiasur.com.ar> +Date: Mon, 28 Dec 2015 17:54:51 -0300 +Subject: [PATCH 086/113] mtd: spi-nor: wait until lock/unlock operations are + ready + +On Micron and Numonyx devices, the status register write command +(WRSR), raises a work-in-progress bit (WIP) on the status register. +The datasheets for these devices specify that while the status +register write is in progress, the status register WIP bit can still +be read to check the end of the operation. + +This commit adds a wait_till_ready call on lock/unlock operations, +which is required for Micron and Numonyx but should be harmless for +others. This is needed to prevent applications from issuing erase or +program operations before the unlock operation is completed. + +Reported-by: Stas Sergeev <stsp@list.ru> +Signed-off-by: Ezequiel Garcia <ezequiel@vanguardiasur.com.ar> +Signed-off-by: Brian Norris <computersforpeace@gmail.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 12 ++++++++++-- + 1 file changed, 10 insertions(+), 2 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -481,6 +481,7 @@ static int stm_lock(struct spi_nor *nor, + int status_old, status_new; + u8 mask = SR_BP2 | SR_BP1 | SR_BP0; + u8 shift = ffs(mask) - 1, pow, val; ++ int ret; + + status_old = read_sr(nor); + if (status_old < 0) +@@ -519,7 +520,10 @@ static int stm_lock(struct spi_nor *nor, + return -EINVAL; + + write_enable(nor); +- return write_sr(nor, status_new); ++ ret = write_sr(nor, status_new); ++ if (ret) ++ return ret; ++ return spi_nor_wait_till_ready(nor); + } + + /* +@@ -533,6 +537,7 @@ static int stm_unlock(struct spi_nor *no + int status_old, status_new; + u8 mask = SR_BP2 | SR_BP1 | SR_BP0; + u8 shift = ffs(mask) - 1, pow, val; ++ int ret; + + status_old = read_sr(nor); + if (status_old < 0) +@@ -569,7 +574,10 @@ static int stm_unlock(struct spi_nor *no + return -EINVAL; + + write_enable(nor); +- return write_sr(nor, status_new); ++ ret = write_sr(nor, status_new); ++ if (ret) ++ return ret; ++ return spi_nor_wait_till_ready(nor); + } + + /* diff --git a/target/linux/layerscape/patches-4.4/1087-mtd-spi-nor-fsl-quadspi-add-big-endian-support.patch b/target/linux/layerscape/patches-4.4/1087-mtd-spi-nor-fsl-quadspi-add-big-endian-support.patch new file mode 100644 index 0000000..9ceddd6 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1087-mtd-spi-nor-fsl-quadspi-add-big-endian-support.patch @@ -0,0 +1,400 @@ +From c58b398221d88ac0db29c3bb7522a4f48dfa102c Mon Sep 17 00:00:00 2001 +From: Yuan Yao <yao.yuan@freescale.com> +Date: Tue, 17 Nov 2015 16:13:47 +0800 +Subject: [PATCH 087/113] mtd: spi-nor: fsl-quadspi: add big-endian support + +Add R/W functions for big- or little-endian registers: +The qSPI controller's endian is independent of the CPU core's endian. +So far, the qSPI have two versions for big-endian and little-endian. + +Signed-off-by: Yuan Yao <yao.yuan@nxp.com> +Acked-by: Han xu <han.xu@freescale.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 157 +++++++++++++++++++++++-------------- + 1 file changed, 97 insertions(+), 60 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -275,6 +275,7 @@ struct fsl_qspi { + u32 clk_rate; + unsigned int chip_base_addr; /* We may support two chips. */ + bool has_second_chip; ++ bool big_endian; + struct mutex lock; + struct pm_qos_request pm_qos_req; + }; +@@ -300,6 +301,28 @@ static inline int needs_wakeup_wait_mode + } + + /* ++ * R/W functions for big- or little-endian registers: ++ * The qSPI controller's endian is independent of the CPU core's endian. ++ * So far, although the CPU core is little-endian but the qSPI have two ++ * versions for big-endian and little-endian. ++ */ ++static void qspi_writel(struct fsl_qspi *q, u32 val, void __iomem *addr) ++{ ++ if (q->big_endian) ++ iowrite32be(val, addr); ++ else ++ iowrite32(val, addr); ++} ++ ++static u32 qspi_readl(struct fsl_qspi *q, void __iomem *addr) ++{ ++ if (q->big_endian) ++ return ioread32be(addr); ++ else ++ return ioread32(addr); ++} ++ ++/* + * An IC bug makes us to re-arrange the 32-bit data. + * The following chips, such as IMX6SLX, have fixed this bug. + */ +@@ -310,14 +333,14 @@ static inline u32 fsl_qspi_endian_xchg(s + + static inline void fsl_qspi_unlock_lut(struct fsl_qspi *q) + { +- writel(QUADSPI_LUTKEY_VALUE, q->iobase + QUADSPI_LUTKEY); +- writel(QUADSPI_LCKER_UNLOCK, q->iobase + QUADSPI_LCKCR); ++ qspi_writel(q, QUADSPI_LUTKEY_VALUE, q->iobase + QUADSPI_LUTKEY); ++ qspi_writel(q, QUADSPI_LCKER_UNLOCK, q->iobase + QUADSPI_LCKCR); + } + + static inline void fsl_qspi_lock_lut(struct fsl_qspi *q) + { +- writel(QUADSPI_LUTKEY_VALUE, q->iobase + QUADSPI_LUTKEY); +- writel(QUADSPI_LCKER_LOCK, q->iobase + QUADSPI_LCKCR); ++ qspi_writel(q, QUADSPI_LUTKEY_VALUE, q->iobase + QUADSPI_LUTKEY); ++ qspi_writel(q, QUADSPI_LCKER_LOCK, q->iobase + QUADSPI_LCKCR); + } + + static irqreturn_t fsl_qspi_irq_handler(int irq, void *dev_id) +@@ -326,8 +349,8 @@ static irqreturn_t fsl_qspi_irq_handler( + u32 reg; + + /* clear interrupt */ +- reg = readl(q->iobase + QUADSPI_FR); +- writel(reg, q->iobase + QUADSPI_FR); ++ reg = qspi_readl(q, q->iobase + QUADSPI_FR); ++ qspi_writel(q, reg, q->iobase + QUADSPI_FR); + + if (reg & QUADSPI_FR_TFF_MASK) + complete(&q->c); +@@ -348,7 +371,7 @@ static void fsl_qspi_init_lut(struct fsl + + /* Clear all the LUT table */ + for (i = 0; i < QUADSPI_LUT_NUM; i++) +- writel(0, base + QUADSPI_LUT_BASE + i * 4); ++ qspi_writel(q, 0, base + QUADSPI_LUT_BASE + i * 4); + + /* Quad Read */ + lut_base = SEQID_QUAD_READ * 4; +@@ -364,14 +387,15 @@ static void fsl_qspi_init_lut(struct fsl + dummy = 8; + } + +- writel(LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), ++ qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); +- writel(LUT0(DUMMY, PAD1, dummy) | LUT1(FSL_READ, PAD4, rxfifo), ++ qspi_writel(q, LUT0(DUMMY, PAD1, dummy) | LUT1(FSL_READ, PAD4, rxfifo), + base + QUADSPI_LUT(lut_base + 1)); + + /* Write enable */ + lut_base = SEQID_WREN * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_WREN), base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_WREN), ++ base + QUADSPI_LUT(lut_base)); + + /* Page Program */ + lut_base = SEQID_PP * 4; +@@ -385,13 +409,15 @@ static void fsl_qspi_init_lut(struct fsl + addrlen = ADDR32BIT; + } + +- writel(LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), ++ qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); +- writel(LUT0(FSL_WRITE, PAD1, 0), base + QUADSPI_LUT(lut_base + 1)); ++ qspi_writel(q, LUT0(FSL_WRITE, PAD1, 0), ++ base + QUADSPI_LUT(lut_base + 1)); + + /* Read Status */ + lut_base = SEQID_RDSR * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_RDSR) | LUT1(FSL_READ, PAD1, 0x1), ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_RDSR) | ++ LUT1(FSL_READ, PAD1, 0x1), + base + QUADSPI_LUT(lut_base)); + + /* Erase a sector */ +@@ -400,40 +426,46 @@ static void fsl_qspi_init_lut(struct fsl + cmd = q->nor[0].erase_opcode; + addrlen = q->nor_size <= SZ_16M ? ADDR24BIT : ADDR32BIT; + +- writel(LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), ++ qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); + + /* Erase the whole chip */ + lut_base = SEQID_CHIP_ERASE * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_CHIP_ERASE), ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_CHIP_ERASE), + base + QUADSPI_LUT(lut_base)); + + /* READ ID */ + lut_base = SEQID_RDID * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_RDID) | LUT1(FSL_READ, PAD1, 0x8), ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_RDID) | ++ LUT1(FSL_READ, PAD1, 0x8), + base + QUADSPI_LUT(lut_base)); + + /* Write Register */ + lut_base = SEQID_WRSR * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_WRSR) | LUT1(FSL_WRITE, PAD1, 0x2), ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_WRSR) | ++ LUT1(FSL_WRITE, PAD1, 0x2), + base + QUADSPI_LUT(lut_base)); + + /* Read Configuration Register */ + lut_base = SEQID_RDCR * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_RDCR) | LUT1(FSL_READ, PAD1, 0x1), ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_RDCR) | ++ LUT1(FSL_READ, PAD1, 0x1), + base + QUADSPI_LUT(lut_base)); + + /* Write disable */ + lut_base = SEQID_WRDI * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_WRDI), base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_WRDI), ++ base + QUADSPI_LUT(lut_base)); + + /* Enter 4 Byte Mode (Micron) */ + lut_base = SEQID_EN4B * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_EN4B), base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_EN4B), ++ base + QUADSPI_LUT(lut_base)); + + /* Enter 4 Byte Mode (Spansion) */ + lut_base = SEQID_BRWR * 4; +- writel(LUT0(CMD, PAD1, SPINOR_OP_BRWR), base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_BRWR), ++ base + QUADSPI_LUT(lut_base)); + + fsl_qspi_lock_lut(q); + } +@@ -488,15 +520,16 @@ fsl_qspi_runcmd(struct fsl_qspi *q, u8 c + q->chip_base_addr, addr, len, cmd); + + /* save the reg */ +- reg = readl(base + QUADSPI_MCR); ++ reg = qspi_readl(q, base + QUADSPI_MCR); + +- writel(q->memmap_phy + q->chip_base_addr + addr, base + QUADSPI_SFAR); +- writel(QUADSPI_RBCT_WMRK_MASK | QUADSPI_RBCT_RXBRD_USEIPS, ++ qspi_writel(q, q->memmap_phy + q->chip_base_addr + addr, ++ base + QUADSPI_SFAR); ++ qspi_writel(q, QUADSPI_RBCT_WMRK_MASK | QUADSPI_RBCT_RXBRD_USEIPS, + base + QUADSPI_RBCT); +- writel(reg | QUADSPI_MCR_CLR_RXF_MASK, base + QUADSPI_MCR); ++ qspi_writel(q, reg | QUADSPI_MCR_CLR_RXF_MASK, base + QUADSPI_MCR); + + do { +- reg2 = readl(base + QUADSPI_SR); ++ reg2 = qspi_readl(q, base + QUADSPI_SR); + if (reg2 & (QUADSPI_SR_IP_ACC_MASK | QUADSPI_SR_AHB_ACC_MASK)) { + udelay(1); + dev_dbg(q->dev, "The controller is busy, 0x%x\n", reg2); +@@ -507,21 +540,22 @@ fsl_qspi_runcmd(struct fsl_qspi *q, u8 c + + /* trigger the LUT now */ + seqid = fsl_qspi_get_seqid(q, cmd); +- writel((seqid << QUADSPI_IPCR_SEQID_SHIFT) | len, base + QUADSPI_IPCR); ++ qspi_writel(q, (seqid << QUADSPI_IPCR_SEQID_SHIFT) | len, ++ base + QUADSPI_IPCR); + + /* Wait for the interrupt. */ + if (!wait_for_completion_timeout(&q->c, msecs_to_jiffies(1000))) { + dev_err(q->dev, + "cmd 0x%.2x timeout, addr@%.8x, FR:0x%.8x, SR:0x%.8x\n", +- cmd, addr, readl(base + QUADSPI_FR), +- readl(base + QUADSPI_SR)); ++ cmd, addr, qspi_readl(q, base + QUADSPI_FR), ++ qspi_readl(q, base + QUADSPI_SR)); + err = -ETIMEDOUT; + } else { + err = 0; + } + + /* restore the MCR */ +- writel(reg, base + QUADSPI_MCR); ++ qspi_writel(q, reg, base + QUADSPI_MCR); + + return err; + } +@@ -533,7 +567,7 @@ static void fsl_qspi_read_data(struct fs + int i = 0; + + while (len > 0) { +- tmp = readl(q->iobase + QUADSPI_RBDR + i * 4); ++ tmp = qspi_readl(q, q->iobase + QUADSPI_RBDR + i * 4); + tmp = fsl_qspi_endian_xchg(q, tmp); + dev_dbg(q->dev, "chip addr:0x%.8x, rcv:0x%.8x\n", + q->chip_base_addr, tmp); +@@ -561,9 +595,9 @@ static inline void fsl_qspi_invalid(stru + { + u32 reg; + +- reg = readl(q->iobase + QUADSPI_MCR); ++ reg = qspi_readl(q, q->iobase + QUADSPI_MCR); + reg |= QUADSPI_MCR_SWRSTHD_MASK | QUADSPI_MCR_SWRSTSD_MASK; +- writel(reg, q->iobase + QUADSPI_MCR); ++ qspi_writel(q, reg, q->iobase + QUADSPI_MCR); + + /* + * The minimum delay : 1 AHB + 2 SFCK clocks. +@@ -572,7 +606,7 @@ static inline void fsl_qspi_invalid(stru + udelay(1); + + reg &= ~(QUADSPI_MCR_SWRSTHD_MASK | QUADSPI_MCR_SWRSTSD_MASK); +- writel(reg, q->iobase + QUADSPI_MCR); ++ qspi_writel(q, reg, q->iobase + QUADSPI_MCR); + } + + static int fsl_qspi_nor_write(struct fsl_qspi *q, struct spi_nor *nor, +@@ -586,20 +620,20 @@ static int fsl_qspi_nor_write(struct fsl + q->chip_base_addr, to, count); + + /* clear the TX FIFO. */ +- tmp = readl(q->iobase + QUADSPI_MCR); +- writel(tmp | QUADSPI_MCR_CLR_TXF_MASK, q->iobase + QUADSPI_MCR); ++ tmp = qspi_readl(q, q->iobase + QUADSPI_MCR); ++ qspi_writel(q, tmp | QUADSPI_MCR_CLR_TXF_MASK, q->iobase + QUADSPI_MCR); + + /* fill the TX data to the FIFO */ + for (j = 0, i = ((count + 3) / 4); j < i; j++) { + tmp = fsl_qspi_endian_xchg(q, *txbuf); +- writel(tmp, q->iobase + QUADSPI_TBDR); ++ qspi_writel(q, tmp, q->iobase + QUADSPI_TBDR); + txbuf++; + } + + /* fill the TXFIFO upto 16 bytes for i.MX7d */ + if (needs_fill_txfifo(q)) + for (; i < 4; i++) +- writel(tmp, q->iobase + QUADSPI_TBDR); ++ qspi_writel(q, tmp, q->iobase + QUADSPI_TBDR); + + /* Trigger it */ + ret = fsl_qspi_runcmd(q, opcode, to, count); +@@ -615,10 +649,10 @@ static void fsl_qspi_set_map_addr(struct + int nor_size = q->nor_size; + void __iomem *base = q->iobase; + +- writel(nor_size + q->memmap_phy, base + QUADSPI_SFA1AD); +- writel(nor_size * 2 + q->memmap_phy, base + QUADSPI_SFA2AD); +- writel(nor_size * 3 + q->memmap_phy, base + QUADSPI_SFB1AD); +- writel(nor_size * 4 + q->memmap_phy, base + QUADSPI_SFB2AD); ++ qspi_writel(q, nor_size + q->memmap_phy, base + QUADSPI_SFA1AD); ++ qspi_writel(q, nor_size * 2 + q->memmap_phy, base + QUADSPI_SFA2AD); ++ qspi_writel(q, nor_size * 3 + q->memmap_phy, base + QUADSPI_SFB1AD); ++ qspi_writel(q, nor_size * 4 + q->memmap_phy, base + QUADSPI_SFB2AD); + } + + /* +@@ -640,24 +674,26 @@ static void fsl_qspi_init_abh_read(struc + int seqid; + + /* AHB configuration for access buffer 0/1/2 .*/ +- writel(QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF0CR); +- writel(QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF1CR); +- writel(QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF2CR); ++ qspi_writel(q, QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF0CR); ++ qspi_writel(q, QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF1CR); ++ qspi_writel(q, QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF2CR); + /* + * Set ADATSZ with the maximum AHB buffer size to improve the + * read performance. + */ +- writel(QUADSPI_BUF3CR_ALLMST_MASK | ((q->devtype_data->ahb_buf_size / 8) +- << QUADSPI_BUF3CR_ADATSZ_SHIFT), base + QUADSPI_BUF3CR); ++ qspi_writel(q, QUADSPI_BUF3CR_ALLMST_MASK | ++ ((q->devtype_data->ahb_buf_size / 8) ++ << QUADSPI_BUF3CR_ADATSZ_SHIFT), ++ base + QUADSPI_BUF3CR); + + /* We only use the buffer3 */ +- writel(0, base + QUADSPI_BUF0IND); +- writel(0, base + QUADSPI_BUF1IND); +- writel(0, base + QUADSPI_BUF2IND); ++ qspi_writel(q, 0, base + QUADSPI_BUF0IND); ++ qspi_writel(q, 0, base + QUADSPI_BUF1IND); ++ qspi_writel(q, 0, base + QUADSPI_BUF2IND); + + /* Set the default lut sequence for AHB Read. */ + seqid = fsl_qspi_get_seqid(q, q->nor[0].read_opcode); +- writel(seqid << QUADSPI_BFGENCR_SEQID_SHIFT, ++ qspi_writel(q, seqid << QUADSPI_BFGENCR_SEQID_SHIFT, + q->iobase + QUADSPI_BFGENCR); + } + +@@ -713,7 +749,7 @@ static int fsl_qspi_nor_setup(struct fsl + return ret; + + /* Reset the module */ +- writel(QUADSPI_MCR_SWRSTSD_MASK | QUADSPI_MCR_SWRSTHD_MASK, ++ qspi_writel(q, QUADSPI_MCR_SWRSTSD_MASK | QUADSPI_MCR_SWRSTHD_MASK, + base + QUADSPI_MCR); + udelay(1); + +@@ -721,24 +757,24 @@ static int fsl_qspi_nor_setup(struct fsl + fsl_qspi_init_lut(q); + + /* Disable the module */ +- writel(QUADSPI_MCR_MDIS_MASK | QUADSPI_MCR_RESERVED_MASK, ++ qspi_writel(q, QUADSPI_MCR_MDIS_MASK | QUADSPI_MCR_RESERVED_MASK, + base + QUADSPI_MCR); + +- reg = readl(base + QUADSPI_SMPR); +- writel(reg & ~(QUADSPI_SMPR_FSDLY_MASK ++ reg = qspi_readl(q, base + QUADSPI_SMPR); ++ qspi_writel(q, reg & ~(QUADSPI_SMPR_FSDLY_MASK + | QUADSPI_SMPR_FSPHS_MASK + | QUADSPI_SMPR_HSENA_MASK + | QUADSPI_SMPR_DDRSMP_MASK), base + QUADSPI_SMPR); + + /* Enable the module */ +- writel(QUADSPI_MCR_RESERVED_MASK | QUADSPI_MCR_END_CFG_MASK, ++ qspi_writel(q, QUADSPI_MCR_RESERVED_MASK | QUADSPI_MCR_END_CFG_MASK, + base + QUADSPI_MCR); + + /* clear all interrupt status */ +- writel(0xffffffff, q->iobase + QUADSPI_FR); ++ qspi_writel(q, 0xffffffff, q->iobase + QUADSPI_FR); + + /* enable the interrupt */ +- writel(QUADSPI_RSER_TFIE, q->iobase + QUADSPI_RSER); ++ qspi_writel(q, QUADSPI_RSER_TFIE, q->iobase + QUADSPI_RSER); + + return 0; + } +@@ -954,6 +990,7 @@ static int fsl_qspi_probe(struct platfor + if (IS_ERR(q->iobase)) + return PTR_ERR(q->iobase); + ++ q->big_endian = of_property_read_bool(np, "big-endian"); + res = platform_get_resource_byname(pdev, IORESOURCE_MEM, + "QuadSPI-memory"); + if (!devm_request_mem_region(dev, res->start, resource_size(res), +@@ -1101,8 +1138,8 @@ static int fsl_qspi_remove(struct platfo + } + + /* disable the hardware */ +- writel(QUADSPI_MCR_MDIS_MASK, q->iobase + QUADSPI_MCR); +- writel(0x0, q->iobase + QUADSPI_RSER); ++ qspi_writel(q, QUADSPI_MCR_MDIS_MASK, q->iobase + QUADSPI_MCR); ++ qspi_writel(q, 0x0, q->iobase + QUADSPI_RSER); + + mutex_destroy(&q->lock); + diff --git a/target/linux/layerscape/patches-4.4/1088-mtd-spi-nor-fsl-quadspi-add-support-for-ls1021a.patch b/target/linux/layerscape/patches-4.4/1088-mtd-spi-nor-fsl-quadspi-add-support-for-ls1021a.patch new file mode 100644 index 0000000..a5782ff --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1088-mtd-spi-nor-fsl-quadspi-add-support-for-ls1021a.patch @@ -0,0 +1,63 @@ +From da44c1517526822e73642fc71b034de8fc7d2b43 Mon Sep 17 00:00:00 2001 +From: Yuan Yao <yao.yuan@freescale.com> +Date: Tue, 17 Nov 2015 16:44:45 +0800 +Subject: [PATCH 088/113] mtd: spi-nor: fsl-quadspi: add support for ls1021a + +[context adjustment] + +LS1021a also support Freescale Quad SPI controller. +Add fsl-quadspi support for ls1021a chip and make SPI_FSL_QUADSPI +selectable for LS1021A SOC hardwares. + +Signed-off-by: Yuan Yao <yao.yuan@nxp.com> +Acked-by: Han xu <han.xu@freescale.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/Kconfig | 2 +- + drivers/mtd/spi-nor/fsl-quadspi.c | 10 ++++++++++ + 2 files changed, 11 insertions(+), 1 deletion(-) + +--- a/drivers/mtd/spi-nor/Kconfig ++++ b/drivers/mtd/spi-nor/Kconfig +@@ -23,7 +23,7 @@ config MTD_SPI_NOR_USE_4K_SECTORS + + config SPI_FSL_QUADSPI + tristate "Freescale Quad SPI controller" +- depends on ARCH_MXC || COMPILE_TEST ++ depends on ARCH_MXC || SOC_LS1021A || COMPILE_TEST + depends on HAS_IOMEM + help + This enables support for the Quad SPI controller in master mode. +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -213,6 +213,7 @@ enum fsl_qspi_devtype { + FSL_QUADSPI_IMX6SX, + FSL_QUADSPI_IMX7D, + FSL_QUADSPI_IMX6UL, ++ FSL_QUADSPI_LS1021A, + }; + + struct fsl_qspi_devtype_data { +@@ -258,6 +259,14 @@ static struct fsl_qspi_devtype_data imx6 + | QUADSPI_QUIRK_4X_INT_CLK, + }; + ++static struct fsl_qspi_devtype_data ls1021a_data = { ++ .devtype = FSL_QUADSPI_LS1021A, ++ .rxfifo = 128, ++ .txfifo = 64, ++ .ahb_buf_size = 1024, ++ .driver_data = 0, ++}; ++ + #define FSL_QSPI_MAX_CHIP 4 + struct fsl_qspi { + struct spi_nor nor[FSL_QSPI_MAX_CHIP]; +@@ -812,6 +821,7 @@ static const struct of_device_id fsl_qsp + { .compatible = "fsl,imx6sx-qspi", .data = (void *)&imx6sx_data, }, + { .compatible = "fsl,imx7d-qspi", .data = (void *)&imx7d_data, }, + { .compatible = "fsl,imx6ul-qspi", .data = (void *)&imx6ul_data, }, ++ { .compatible = "fsl,ls1021a-qspi", .data = (void *)&ls1021a_data, }, + { /* sentinel */ } + }; + MODULE_DEVICE_TABLE(of, fsl_qspi_dt_ids); diff --git a/target/linux/layerscape/patches-4.4/1089-mtd-spi-nor-fsl-quadspi-add-support-for-layerscape.patch b/target/linux/layerscape/patches-4.4/1089-mtd-spi-nor-fsl-quadspi-add-support-for-layerscape.patch new file mode 100644 index 0000000..bd959df --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1089-mtd-spi-nor-fsl-quadspi-add-support-for-layerscape.patch @@ -0,0 +1,28 @@ +From 9c6153130081ef2c109e2a243a598f2bc0dc6413 Mon Sep 17 00:00:00 2001 +From: Yuan Yao <yao.yuan@freescale.com> +Date: Tue, 17 Nov 2015 17:06:47 +0800 +Subject: [PATCH 089/113] mtd: spi-nor: fsl-quadspi: add support for + layerscape + +[context adjustment] + +LS1043a and LS2080A in the Layerscape family also support Freescale Quad +SPI, make Quad SPI selectable for these hardwares. + +Signed-off-by: Yuan Yao <yao.yuan@nxp.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/Kconfig | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +--- a/drivers/mtd/spi-nor/Kconfig ++++ b/drivers/mtd/spi-nor/Kconfig +@@ -23,7 +23,7 @@ config MTD_SPI_NOR_USE_4K_SECTORS + + config SPI_FSL_QUADSPI + tristate "Freescale Quad SPI controller" +- depends on ARCH_MXC || SOC_LS1021A || COMPILE_TEST ++ depends on ARCH_MXC || SOC_LS1021A || ARCH_LAYERSCAPE || COMPILE_TEST + depends on HAS_IOMEM + help + This enables support for the Quad SPI controller in master mode. diff --git a/target/linux/layerscape/patches-4.4/1090-mtd-spi-nor-Add-SPI-NOR-layer-PM-support.patch b/target/linux/layerscape/patches-4.4/1090-mtd-spi-nor-Add-SPI-NOR-layer-PM-support.patch new file mode 100644 index 0000000..e959ca5 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1090-mtd-spi-nor-Add-SPI-NOR-layer-PM-support.patch @@ -0,0 +1,137 @@ +From 2c5a3db21926e9ebfd7a32e3c36a3256ed84903c Mon Sep 17 00:00:00 2001 +From: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +Date: Thu, 19 Nov 2015 20:25:24 +0800 +Subject: [PATCH 090/113] mtd: spi-nor: Add SPI NOR layer PM support + +[context adjustment] + +Add the Power Management API in SPI NOR framework. +The Power Management system will turn off power supply to SPI flash +when system suspending, and then the SPI flash will be in the reset +state after system resuming. As a result, the status&configurations +of SPI flash driver will mismatch with its current hardware state. +So reinitialize SPI flash to make sure it is resumed to the correct +state. +And the SPI NOR layer just do common configuration depending on the +records in structure spi_nor. + +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 74 ++++++++++++++++++++++++++++++++++------- + include/linux/mtd/spi-nor.h | 9 +++++ + 2 files changed, 71 insertions(+), 12 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -1139,6 +1139,26 @@ static int spi_nor_check(struct spi_nor + return 0; + } + ++/* ++ * Atmel, SST, Intel/Numonyx, and others serial NOR tend to power up ++ * with the software protection bits set ++ */ ++static int spi_nor_unprotect_on_powerup(struct spi_nor *nor) ++{ ++ const struct flash_info *info = NULL; ++ int ret = 0; ++ ++ info = spi_nor_read_id(nor); ++ if (JEDEC_MFR(info) == SNOR_MFR_ATMEL || ++ JEDEC_MFR(info) == SNOR_MFR_INTEL || ++ JEDEC_MFR(info) == SNOR_MFR_SST) { ++ write_enable(nor); ++ ret = write_sr(nor, 0); ++ } ++ ++ return ret; ++} ++ + int spi_nor_scan(struct spi_nor *nor, const char *name, enum read_mode mode) + { + const struct flash_info *info = NULL; +@@ -1186,18 +1206,9 @@ int spi_nor_scan(struct spi_nor *nor, co + + mutex_init(&nor->lock); + +- /* +- * Atmel, SST, Intel/Numonyx, and others serial NOR tend to power up +- * with the software protection bits set +- */ +- +- if (JEDEC_MFR(info) == SNOR_MFR_ATMEL || +- JEDEC_MFR(info) == SNOR_MFR_INTEL || +- JEDEC_MFR(info) == SNOR_MFR_MACRONIX || +- JEDEC_MFR(info) == SNOR_MFR_SST) { +- write_enable(nor); +- write_sr(nor, 0); +- } ++ ret = spi_nor_unprotect_on_powerup(nor); ++ if (ret) ++ return ret; + + if (!mtd->name) + mtd->name = dev_name(dev); +@@ -1363,6 +1374,45 @@ int spi_nor_scan(struct spi_nor *nor, co + } + EXPORT_SYMBOL_GPL(spi_nor_scan); + ++static int spi_nor_hw_reinit(struct spi_nor *nor) ++{ ++ const struct flash_info *info = NULL; ++ struct device *dev = nor->dev; ++ int ret; ++ ++ info = spi_nor_read_id(nor); ++ ++ ret = spi_nor_unprotect_on_powerup(nor); ++ if (ret) ++ return ret; ++ ++ if (nor->flash_read == SPI_NOR_QUAD) { ++ ret = set_quad_mode(nor, info); ++ if (ret) { ++ dev_err(dev, "quad mode not supported\n"); ++ return ret; ++ } ++ } ++ ++ if (nor->addr_width == 4 && ++ JEDEC_MFR(info) != SNOR_MFR_SPANSION) ++ set_4byte(nor, info, 1); ++ ++ return 0; ++} ++ ++int spi_nor_suspend(struct spi_nor *nor) ++{ ++ return 0; ++} ++EXPORT_SYMBOL_GPL(spi_nor_suspend); ++ ++int spi_nor_resume(struct spi_nor *nor) ++{ ++ return spi_nor_hw_reinit(nor); ++} ++EXPORT_SYMBOL_GPL(spi_nor_resume); ++ + static const struct flash_info *spi_nor_match_id(const char *name) + { + const struct flash_info *id = spi_nor_ids; +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -210,4 +210,13 @@ static inline struct device_node *spi_no + */ + int spi_nor_scan(struct spi_nor *nor, const char *name, enum read_mode mode); + ++/** ++ * spi_nor_suspend/resume() - the SPI NOR layer PM API ++ * @nor: the spi_nor structure ++ * ++ * Return: 0 for success, others for failure. ++ */ ++int spi_nor_suspend(struct spi_nor *nor); ++int spi_nor_resume(struct spi_nor *nor); ++ + #endif diff --git a/target/linux/layerscape/patches-4.4/1091-mtd-spi-nor-change-return-value-of-read-write.patch b/target/linux/layerscape/patches-4.4/1091-mtd-spi-nor-change-return-value-of-read-write.patch new file mode 100644 index 0000000..faa806e --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1091-mtd-spi-nor-change-return-value-of-read-write.patch @@ -0,0 +1,82 @@ +From 0a8079b232e9188ba267e37e20f192bed6c2b29b Mon Sep 17 00:00:00 2001 +From: Michal Suchanek <hramrach@gmail.com> +Date: Wed, 2 Dec 2015 10:38:19 +0000 +Subject: [PATCH 091/113] mtd: spi-nor: change return value of read/write + +Change the return value of spi-nor device read and write methods to +allow returning amount of data transferred and errors as +read(2)/write(2) does. + +Signed-off-by: Michal Suchanek <hramrach@gmail.com> +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +--- + drivers/mtd/devices/m25p80.c | 5 +++-- + drivers/mtd/spi-nor/fsl-quadspi.c | 5 +++-- + include/linux/mtd/spi-nor.h | 4 ++-- + 3 files changed, 8 insertions(+), 6 deletions(-) + +--- a/drivers/mtd/devices/m25p80.c ++++ b/drivers/mtd/devices/m25p80.c +@@ -73,7 +73,7 @@ static int m25p80_write_reg(struct spi_n + return spi_write(spi, flash->command, len + 1); + } + +-static void m25p80_write(struct spi_nor *nor, loff_t to, size_t len, ++static ssize_t m25p80_write(struct spi_nor *nor, loff_t to, size_t len, + size_t *retlen, const u_char *buf) + { + struct m25p *flash = nor->priv; +@@ -101,6 +101,7 @@ static void m25p80_write(struct spi_nor + spi_sync(spi, &m); + + *retlen += m.actual_length - cmd_sz; ++ return 0; + } + + static inline unsigned int m25p80_rx_nbits(struct spi_nor *nor) +@@ -119,7 +120,7 @@ static inline unsigned int m25p80_rx_nbi + * Read an address range from the nor chip. The address range + * may be any size provided it is within the physical boundaries. + */ +-static int m25p80_read(struct spi_nor *nor, loff_t from, size_t len, ++static ssize_t m25p80_read(struct spi_nor *nor, loff_t from, size_t len, + size_t *retlen, u_char *buf) + { + struct m25p *flash = nor->priv; +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -868,7 +868,7 @@ static int fsl_qspi_write_reg(struct spi + return ret; + } + +-static void fsl_qspi_write(struct spi_nor *nor, loff_t to, ++static ssize_t fsl_qspi_write(struct spi_nor *nor, loff_t to, + size_t len, size_t *retlen, const u_char *buf) + { + struct fsl_qspi *q = nor->priv; +@@ -878,9 +878,10 @@ static void fsl_qspi_write(struct spi_no + + /* invalid the data in the AHB buffer. */ + fsl_qspi_invalid(q); ++ return 0; + } + +-static int fsl_qspi_read(struct spi_nor *nor, loff_t from, ++static ssize_t fsl_qspi_read(struct spi_nor *nor, loff_t from, + size_t len, size_t *retlen, u_char *buf) + { + struct fsl_qspi *q = nor->priv; +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -170,9 +170,9 @@ struct spi_nor { + int (*read_reg)(struct spi_nor *nor, u8 opcode, u8 *buf, int len); + int (*write_reg)(struct spi_nor *nor, u8 opcode, u8 *buf, int len); + +- int (*read)(struct spi_nor *nor, loff_t from, ++ ssize_t (*read)(struct spi_nor *nor, loff_t from, + size_t len, size_t *retlen, u_char *read_buf); +- void (*write)(struct spi_nor *nor, loff_t to, ++ ssize_t (*write)(struct spi_nor *nor, loff_t to, + size_t len, size_t *retlen, const u_char *write_buf); + int (*erase)(struct spi_nor *nor, loff_t offs); + diff --git a/target/linux/layerscape/patches-4.4/1092-mtd-fsl-quadspi-return-amount-of-data-read-written-o.patch b/target/linux/layerscape/patches-4.4/1092-mtd-fsl-quadspi-return-amount-of-data-read-written-o.patch new file mode 100644 index 0000000..4438036 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1092-mtd-fsl-quadspi-return-amount-of-data-read-written-o.patch @@ -0,0 +1,73 @@ +From 99768b3062501b05810fb62545279da3a4371ca0 Mon Sep 17 00:00:00 2001 +From: Michal Suchanek <hramrach@gmail.com> +Date: Wed, 2 Dec 2015 10:38:19 +0000 +Subject: [PATCH 092/113] mtd: fsl-quadspi: return amount of data read/written + or error + +Return amount of data read/written or error as read(2)/write(2) does. + +Signed-off-by: Michal Suchanek <hramrach@gmail.com> +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 18 +++++++++++------- + 1 file changed, 11 insertions(+), 7 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -618,7 +618,7 @@ static inline void fsl_qspi_invalid(stru + qspi_writel(q, reg, q->iobase + QUADSPI_MCR); + } + +-static int fsl_qspi_nor_write(struct fsl_qspi *q, struct spi_nor *nor, ++static ssize_t fsl_qspi_nor_write(struct fsl_qspi *q, struct spi_nor *nor, + u8 opcode, unsigned int to, u32 *txbuf, + unsigned count, size_t *retlen) + { +@@ -647,8 +647,11 @@ static int fsl_qspi_nor_write(struct fsl + /* Trigger it */ + ret = fsl_qspi_runcmd(q, opcode, to, count); + +- if (ret == 0 && retlen) +- *retlen += count; ++ if (ret == 0) { ++ if (retlen) ++ *retlen += count; ++ return count; ++ } + + return ret; + } +@@ -860,6 +863,8 @@ static int fsl_qspi_write_reg(struct spi + } else if (len > 0) { + ret = fsl_qspi_nor_write(q, nor, opcode, 0, + (u32 *)buf, len, NULL); ++ if (ret > 0) ++ return 0; + } else { + dev_err(q->dev, "invalid cmd %d\n", opcode); + ret = -EINVAL; +@@ -873,12 +878,12 @@ static ssize_t fsl_qspi_write(struct spi + { + struct fsl_qspi *q = nor->priv; + +- fsl_qspi_nor_write(q, nor, nor->program_opcode, to, ++ ssize_t ret = fsl_qspi_nor_write(q, nor, nor->program_opcode, to, + (u32 *)buf, len, retlen); + + /* invalid the data in the AHB buffer. */ + fsl_qspi_invalid(q); +- return 0; ++ return ret; + } + + static ssize_t fsl_qspi_read(struct spi_nor *nor, loff_t from, +@@ -924,8 +929,7 @@ static ssize_t fsl_qspi_read(struct spi_ + memcpy(buf, q->ahb_addr + q->chip_base_addr + from - q->memmap_offs, + len); + +- *retlen += len; +- return 0; ++ return len; + } + + static int fsl_qspi_erase(struct spi_nor *nor, loff_t offs) diff --git a/target/linux/layerscape/patches-4.4/1093-mtd-spi-nor-check-return-value-from-read-write.patch b/target/linux/layerscape/patches-4.4/1093-mtd-spi-nor-check-return-value-from-read-write.patch new file mode 100644 index 0000000..0721a06 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1093-mtd-spi-nor-check-return-value-from-read-write.patch @@ -0,0 +1,127 @@ +From 8527843351169d999995d331bbdad75560ccafb2 Mon Sep 17 00:00:00 2001 +From: Michal Suchanek <hramrach@gmail.com> +Date: Wed, 2 Dec 2015 10:38:20 +0000 +Subject: [PATCH 093/113] mtd: spi-nor: check return value from read/write + +SPI NOR hardware drivers now return useful value from their read/write +functions so check them. + +Signed-off-by: Michal Suchanek <hramrach@gmail.com> +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 50 +++++++++++++++++++++++++++++------------ + 1 file changed, 36 insertions(+), 14 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -922,7 +922,10 @@ static int spi_nor_read(struct mtd_info + ret = nor->read(nor, from, len, retlen, buf); + + spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_READ); +- return ret; ++ if (ret < 0) ++ return ret; ++ ++ return 0; + } + + static int sst_write(struct mtd_info *mtd, loff_t to, size_t len, +@@ -948,10 +951,14 @@ static int sst_write(struct mtd_info *mt + nor->program_opcode = SPINOR_OP_BP; + + /* write one byte. */ +- nor->write(nor, to, 1, retlen, buf); ++ ret = nor->write(nor, to, 1, retlen, buf); ++ if (ret < 0) ++ goto sst_write_err; ++ WARN(ret != 1, "While writing 1 byte written %i bytes\n", ++ (int)ret); + ret = spi_nor_wait_till_ready(nor); + if (ret) +- goto time_out; ++ goto sst_write_err; + } + to += actual; + +@@ -960,10 +967,14 @@ static int sst_write(struct mtd_info *mt + nor->program_opcode = SPINOR_OP_AAI_WP; + + /* write two bytes. */ +- nor->write(nor, to, 2, retlen, buf + actual); ++ ret = nor->write(nor, to, 2, retlen, buf + actual); ++ if (ret < 0) ++ goto sst_write_err; ++ WARN(ret != 2, "While writing 2 bytes written %i bytes\n", ++ (int)ret); + ret = spi_nor_wait_till_ready(nor); + if (ret) +- goto time_out; ++ goto sst_write_err; + to += 2; + nor->sst_write_second = true; + } +@@ -972,21 +983,24 @@ static int sst_write(struct mtd_info *mt + write_disable(nor); + ret = spi_nor_wait_till_ready(nor); + if (ret) +- goto time_out; ++ goto sst_write_err; + + /* Write out trailing byte if it exists. */ + if (actual != len) { + write_enable(nor); + + nor->program_opcode = SPINOR_OP_BP; +- nor->write(nor, to, 1, retlen, buf + actual); +- ++ ret = nor->write(nor, to, 1, retlen, buf + actual); ++ if (ret < 0) ++ goto sst_write_err; ++ WARN(ret != 1, "While writing 1 byte written %i bytes\n", ++ (int)ret); + ret = spi_nor_wait_till_ready(nor); + if (ret) +- goto time_out; ++ goto sst_write_err; + write_disable(nor); + } +-time_out: ++sst_write_err: + spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_WRITE); + return ret; + } +@@ -1015,14 +1029,18 @@ static int spi_nor_write(struct mtd_info + + /* do all the bytes fit onto one page? */ + if (page_offset + len <= nor->page_size) { +- nor->write(nor, to, len, retlen, buf); ++ ret = nor->write(nor, to, len, retlen, buf); ++ if (ret < 0) ++ goto write_err; + } else { + /* the size of data remaining on the first page */ + page_size = nor->page_size - page_offset; +- nor->write(nor, to, page_size, retlen, buf); ++ ret = nor->write(nor, to, page_size, retlen, buf); ++ if (ret < 0) ++ goto write_err; + + /* write everything in nor->page_size chunks */ +- for (i = page_size; i < len; i += page_size) { ++ for (i = ret; i < len; ) { + page_size = len - i; + if (page_size > nor->page_size) + page_size = nor->page_size; +@@ -1033,7 +1051,11 @@ static int spi_nor_write(struct mtd_info + + write_enable(nor); + +- nor->write(nor, to + i, page_size, retlen, buf + i); ++ ret = nor->write(nor, to + i, page_size, retlen, ++ buf + i); ++ if (ret < 0) ++ goto write_err; ++ i += ret; + } + } + diff --git a/target/linux/layerscape/patches-4.4/1094-mtd-spi-nor-stop-passing-around-retlen.patch b/target/linux/layerscape/patches-4.4/1094-mtd-spi-nor-stop-passing-around-retlen.patch new file mode 100644 index 0000000..6223dfc --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1094-mtd-spi-nor-stop-passing-around-retlen.patch @@ -0,0 +1,215 @@ +From a99477d72b500b48cb3614aad0ce096fe4e3f437 Mon Sep 17 00:00:00 2001 +From: Michal Suchanek <hramrach@gmail.com> +Date: Wed, 2 Dec 2015 10:38:20 +0000 +Subject: [PATCH 094/113] mtd: spi-nor: stop passing around retlen + +[context adjustment] +not apply changes of drivers/mtd/devices/m25p80.c +################# +@@ -74,7 +74,7 @@ static int m25p80_write_reg(struct spi_nor *nor, u8 opcode, u8 *buf, int len) + } + + static ssize_t m25p80_write(struct spi_nor *nor, loff_t to, size_t len, +- size_t *retlen, const u_char *buf) ++ const u_char *buf) + { + struct m25p *flash = nor->priv; + struct spi_device *spi = flash->spi; +@@ -106,7 +106,6 @@ static ssize_t m25p80_write(struct spi_nor *nor, loff_t to, size_t len, + ret = m.actual_length - cmd_sz; + if (ret < 0) + return -EIO; +- *retlen += ret; + return ret; + } + +@@ -127,7 +126,7 @@ static inline unsigned int m25p80_rx_nbits(struct spi_nor *nor) + * may be any size provided it is within the physical boundaries. + */ + static ssize_t m25p80_read(struct spi_nor *nor, loff_t from, size_t len, +- size_t *retlen, u_char *buf) ++ u_char *buf) + { + struct m25p *flash = nor->priv; + struct spi_device *spi = flash->spi; +@@ -161,7 +160,6 @@ static ssize_t m25p80_read(struct spi_nor *nor, loff_t from, size_t len, + ret = m.actual_length - m25p_cmdsz(nor) - dummy; + if (ret < 0) + return -EIO; +- *retlen += ret; + return ret; + } + +################# + +Do not pass retlen to hardware driver read/write functions. Update it in +spi-nor generic driver instead. + +Signed-off-by: Michal Suchanek <hramrach@gmail.com> +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 16 ++++++---------- + drivers/mtd/spi-nor/spi-nor.c | 21 +++++++++++++-------- + include/linux/mtd/spi-nor.h | 4 ++-- + 3 files changed, 21 insertions(+), 20 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -620,7 +620,7 @@ static inline void fsl_qspi_invalid(stru + + static ssize_t fsl_qspi_nor_write(struct fsl_qspi *q, struct spi_nor *nor, + u8 opcode, unsigned int to, u32 *txbuf, +- unsigned count, size_t *retlen) ++ unsigned count) + { + int ret, i, j; + u32 tmp; +@@ -647,11 +647,8 @@ static ssize_t fsl_qspi_nor_write(struct + /* Trigger it */ + ret = fsl_qspi_runcmd(q, opcode, to, count); + +- if (ret == 0) { +- if (retlen) +- *retlen += count; ++ if (ret == 0) + return count; +- } + + return ret; + } +@@ -862,7 +859,7 @@ static int fsl_qspi_write_reg(struct spi + + } else if (len > 0) { + ret = fsl_qspi_nor_write(q, nor, opcode, 0, +- (u32 *)buf, len, NULL); ++ (u32 *)buf, len); + if (ret > 0) + return 0; + } else { +@@ -874,12 +871,11 @@ static int fsl_qspi_write_reg(struct spi + } + + static ssize_t fsl_qspi_write(struct spi_nor *nor, loff_t to, +- size_t len, size_t *retlen, const u_char *buf) ++ size_t len, const u_char *buf) + { + struct fsl_qspi *q = nor->priv; +- + ssize_t ret = fsl_qspi_nor_write(q, nor, nor->program_opcode, to, +- (u32 *)buf, len, retlen); ++ (u32 *)buf, len); + + /* invalid the data in the AHB buffer. */ + fsl_qspi_invalid(q); +@@ -887,7 +883,7 @@ static ssize_t fsl_qspi_write(struct spi + } + + static ssize_t fsl_qspi_read(struct spi_nor *nor, loff_t from, +- size_t len, size_t *retlen, u_char *buf) ++ size_t len, u_char *buf) + { + struct fsl_qspi *q = nor->priv; + u8 cmd = nor->read_opcode; +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -919,12 +919,13 @@ static int spi_nor_read(struct mtd_info + if (ret) + return ret; + +- ret = nor->read(nor, from, len, retlen, buf); ++ ret = nor->read(nor, from, len, buf); + + spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_READ); + if (ret < 0) + return ret; + ++ *retlen += ret; + return 0; + } + +@@ -951,7 +952,7 @@ static int sst_write(struct mtd_info *mt + nor->program_opcode = SPINOR_OP_BP; + + /* write one byte. */ +- ret = nor->write(nor, to, 1, retlen, buf); ++ ret = nor->write(nor, to, 1, buf); + if (ret < 0) + goto sst_write_err; + WARN(ret != 1, "While writing 1 byte written %i bytes\n", +@@ -967,7 +968,7 @@ static int sst_write(struct mtd_info *mt + nor->program_opcode = SPINOR_OP_AAI_WP; + + /* write two bytes. */ +- ret = nor->write(nor, to, 2, retlen, buf + actual); ++ ret = nor->write(nor, to, 2, buf + actual); + if (ret < 0) + goto sst_write_err; + WARN(ret != 2, "While writing 2 bytes written %i bytes\n", +@@ -990,7 +991,7 @@ static int sst_write(struct mtd_info *mt + write_enable(nor); + + nor->program_opcode = SPINOR_OP_BP; +- ret = nor->write(nor, to, 1, retlen, buf + actual); ++ ret = nor->write(nor, to, 1, buf + actual); + if (ret < 0) + goto sst_write_err; + WARN(ret != 1, "While writing 1 byte written %i bytes\n", +@@ -999,8 +1000,10 @@ static int sst_write(struct mtd_info *mt + if (ret) + goto sst_write_err; + write_disable(nor); ++ actual += 1; + } + sst_write_err: ++ *retlen += actual; + spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_WRITE); + return ret; + } +@@ -1029,15 +1032,17 @@ static int spi_nor_write(struct mtd_info + + /* do all the bytes fit onto one page? */ + if (page_offset + len <= nor->page_size) { +- ret = nor->write(nor, to, len, retlen, buf); ++ ret = nor->write(nor, to, len, buf); + if (ret < 0) + goto write_err; ++ *retlen += ret; + } else { + /* the size of data remaining on the first page */ + page_size = nor->page_size - page_offset; +- ret = nor->write(nor, to, page_size, retlen, buf); ++ ret = nor->write(nor, to, page_size, buf); + if (ret < 0) + goto write_err; ++ *retlen += ret; + + /* write everything in nor->page_size chunks */ + for (i = ret; i < len; ) { +@@ -1051,10 +1056,10 @@ static int spi_nor_write(struct mtd_info + + write_enable(nor); + +- ret = nor->write(nor, to + i, page_size, retlen, +- buf + i); ++ ret = nor->write(nor, to + i, page_size, buf + i); + if (ret < 0) + goto write_err; ++ *retlen += ret; + i += ret; + } + } +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -171,9 +171,9 @@ struct spi_nor { + int (*write_reg)(struct spi_nor *nor, u8 opcode, u8 *buf, int len); + + ssize_t (*read)(struct spi_nor *nor, loff_t from, +- size_t len, size_t *retlen, u_char *read_buf); ++ size_t len, u_char *read_buf); + ssize_t (*write)(struct spi_nor *nor, loff_t to, +- size_t len, size_t *retlen, const u_char *write_buf); ++ size_t len, const u_char *write_buf); + int (*erase)(struct spi_nor *nor, loff_t offs); + + int (*flash_lock)(struct spi_nor *nor, loff_t ofs, uint64_t len); diff --git a/target/linux/layerscape/patches-4.4/1095-mtd-spi-nor-simplify-write-loop.patch b/target/linux/layerscape/patches-4.4/1095-mtd-spi-nor-simplify-write-loop.patch new file mode 100644 index 0000000..8eb957d --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1095-mtd-spi-nor-simplify-write-loop.patch @@ -0,0 +1,100 @@ +From 93b40e12f7e580a41c4aee5597579cc539fd8544 Mon Sep 17 00:00:00 2001 +From: Michal Suchanek <hramrach@gmail.com> +Date: Wed, 2 Dec 2015 10:38:20 +0000 +Subject: [PATCH 095/113] mtd: spi-nor: simplify write loop + +The spi-nor write loop assumes that what is passed to the hardware +driver write() is what gets written. + +When write() writes less than page size at once data is dropped on the +floor. Check the amount of data writen and exit if it does not match +requested amount. + +Signed-off-by: Michal Suchanek <hramrach@gmail.com> +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 58 ++++++++++++++++++----------------------- + 1 file changed, 25 insertions(+), 33 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -1017,8 +1017,8 @@ static int spi_nor_write(struct mtd_info + size_t *retlen, const u_char *buf) + { + struct spi_nor *nor = mtd_to_spi_nor(mtd); +- u32 page_offset, page_size, i; +- int ret; ++ size_t page_offset, page_remain, i; ++ ssize_t ret; + + dev_dbg(nor->dev, "to 0x%08x, len %zd\n", (u32)to, len); + +@@ -1026,45 +1026,37 @@ static int spi_nor_write(struct mtd_info + if (ret) + return ret; + +- write_enable(nor); ++ for (i = 0; i < len; ) { ++ ssize_t written; + +- page_offset = to & (nor->page_size - 1); +- +- /* do all the bytes fit onto one page? */ +- if (page_offset + len <= nor->page_size) { +- ret = nor->write(nor, to, len, buf); +- if (ret < 0) +- goto write_err; +- *retlen += ret; +- } else { ++ page_offset = to & (nor->page_size - 1); ++ WARN_ONCE(page_offset, ++ "Writing at offset %zu into a NOR page. Writing partial pages may decrease reliability and increase wear of NOR flash.", ++ page_offset); + /* the size of data remaining on the first page */ +- page_size = nor->page_size - page_offset; +- ret = nor->write(nor, to, page_size, buf); ++ page_remain = min_t(size_t, ++ nor->page_size - page_offset, len - i); ++ ++ write_enable(nor); ++ ret = nor->write(nor, to + i, page_remain, buf + i); + if (ret < 0) + goto write_err; +- *retlen += ret; ++ written = ret; + +- /* write everything in nor->page_size chunks */ +- for (i = ret; i < len; ) { +- page_size = len - i; +- if (page_size > nor->page_size) +- page_size = nor->page_size; +- +- ret = spi_nor_wait_till_ready(nor); +- if (ret) +- goto write_err; +- +- write_enable(nor); +- +- ret = nor->write(nor, to + i, page_size, buf + i); +- if (ret < 0) +- goto write_err; +- *retlen += ret; +- i += ret; ++ ret = spi_nor_wait_till_ready(nor); ++ if (ret) ++ goto write_err; ++ *retlen += written; ++ i += written; ++ if (written != page_remain) { ++ dev_err(nor->dev, ++ "While writing %zu bytes written %zd bytes\n", ++ page_remain, written); ++ ret = -EIO; ++ goto write_err; + } + } + +- ret = spi_nor_wait_till_ready(nor); + write_err: + spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_WRITE); + return ret; diff --git a/target/linux/layerscape/patches-4.4/1096-mtd-spi-nor-add-read-loop.patch b/target/linux/layerscape/patches-4.4/1096-mtd-spi-nor-add-read-loop.patch new file mode 100644 index 0000000..db557e9 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1096-mtd-spi-nor-add-read-loop.patch @@ -0,0 +1,46 @@ +From b5929f91416d64afacf46c649f38cc8f0eea50d2 Mon Sep 17 00:00:00 2001 +From: Michal Suchanek <hramrach@gmail.com> +Date: Wed, 2 Dec 2015 10:38:20 +0000 +Subject: [PATCH 096/113] mtd: spi-nor: add read loop + +mtdblock and ubi do not handle the situation when read returns less data +than requested. Loop in spi-nor until buffer is filled or an error is +returned. + +Signed-off-by: Michal Suchanek <hramrach@gmail.com> +Signed-off-by: Hou Zhiqiang <Zhiqiang.Hou@freescale.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 20 ++++++++++++++------ + 1 file changed, 14 insertions(+), 6 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -919,14 +919,22 @@ static int spi_nor_read(struct mtd_info + if (ret) + return ret; + +- ret = nor->read(nor, from, len, buf); ++ while (len) { ++ ret = nor->read(nor, from, len, buf); ++ if (ret <= 0) ++ goto read_err; + +- spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_READ); +- if (ret < 0) +- return ret; ++ WARN_ON(ret > len); ++ *retlen += ret; ++ buf += ret; ++ from += ret; ++ len -= ret; ++ } ++ ret = 0; + +- *retlen += ret; +- return 0; ++read_err: ++ spi_nor_unlock_and_unprep(nor, SPI_NOR_OPS_READ); ++ return ret; + } + + static int sst_write(struct mtd_info *mtd, loff_t to, size_t len, diff --git a/target/linux/layerscape/patches-4.4/1097-mtd-fsl-quadspi-use-the-property-fields-of-SPI-NOR.patch b/target/linux/layerscape/patches-4.4/1097-mtd-fsl-quadspi-use-the-property-fields-of-SPI-NOR.patch new file mode 100644 index 0000000..d07264d --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1097-mtd-fsl-quadspi-use-the-property-fields-of-SPI-NOR.patch @@ -0,0 +1,87 @@ +From 5c315652c1b43a6a3abe48c2842cde822ac0ff3c Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <B56489@freescale.com> +Date: Wed, 20 Jan 2016 18:40:31 +0800 +Subject: [PATCH 097/113] mtd:fsl-quadspi:use the property fields of SPI-NOR + +We can get the read/write/erase opcode from the spi nor framework +directly. This patch uses the information stored in the SPI-NOR to +remove the hardcode in the fsl_qspi_init_lut(). + +Signed-off-by: Yunhui Cui <B56489@freescale.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 40 +++++++++++-------------------------- + 1 file changed, 12 insertions(+), 28 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -373,9 +373,13 @@ static void fsl_qspi_init_lut(struct fsl + void __iomem *base = q->iobase; + int rxfifo = q->devtype_data->rxfifo; + u32 lut_base; +- u8 cmd, addrlen, dummy; + int i; + ++ struct spi_nor *nor = &q->nor[0]; ++ u8 addrlen = (nor->addr_width == 3) ? ADDR24BIT : ADDR32BIT; ++ u8 read_op = nor->read_opcode; ++ u8 read_dm = nor->read_dummy; ++ + fsl_qspi_unlock_lut(q); + + /* Clear all the LUT table */ +@@ -385,20 +389,10 @@ static void fsl_qspi_init_lut(struct fsl + /* Quad Read */ + lut_base = SEQID_QUAD_READ * 4; + +- if (q->nor_size <= SZ_16M) { +- cmd = SPINOR_OP_READ_1_1_4; +- addrlen = ADDR24BIT; +- dummy = 8; +- } else { +- /* use the 4-byte address */ +- cmd = SPINOR_OP_READ_1_1_4; +- addrlen = ADDR32BIT; +- dummy = 8; +- } +- +- qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), ++ qspi_writel(q, LUT0(CMD, PAD1, read_op) | LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); +- qspi_writel(q, LUT0(DUMMY, PAD1, dummy) | LUT1(FSL_READ, PAD4, rxfifo), ++ qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | ++ LUT1(FSL_READ, PAD4, rxfifo), + base + QUADSPI_LUT(lut_base + 1)); + + /* Write enable */ +@@ -409,16 +403,8 @@ static void fsl_qspi_init_lut(struct fsl + /* Page Program */ + lut_base = SEQID_PP * 4; + +- if (q->nor_size <= SZ_16M) { +- cmd = SPINOR_OP_PP; +- addrlen = ADDR24BIT; +- } else { +- /* use the 4-byte address */ +- cmd = SPINOR_OP_PP; +- addrlen = ADDR32BIT; +- } +- +- qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), ++ qspi_writel(q, LUT0(CMD, PAD1, nor->program_opcode) | ++ LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); + qspi_writel(q, LUT0(FSL_WRITE, PAD1, 0), + base + QUADSPI_LUT(lut_base + 1)); +@@ -432,10 +418,8 @@ static void fsl_qspi_init_lut(struct fsl + /* Erase a sector */ + lut_base = SEQID_SE * 4; + +- cmd = q->nor[0].erase_opcode; +- addrlen = q->nor_size <= SZ_16M ? ADDR24BIT : ADDR32BIT; +- +- qspi_writel(q, LUT0(CMD, PAD1, cmd) | LUT1(ADDR, PAD1, addrlen), ++ qspi_writel(q, LUT0(CMD, PAD1, nor->erase_opcode) | ++ LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); + + /* Erase the whole chip */ diff --git a/target/linux/layerscape/patches-4.4/1098-mtd-fsl-quadspi-Rename-SEQID_QUAD_READ-to-SEQID_READ.patch b/target/linux/layerscape/patches-4.4/1098-mtd-fsl-quadspi-Rename-SEQID_QUAD_READ-to-SEQID_READ.patch new file mode 100644 index 0000000..f27ea21 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1098-mtd-fsl-quadspi-Rename-SEQID_QUAD_READ-to-SEQID_READ.patch @@ -0,0 +1,46 @@ +From c8f9be7df954fce18e96074af3f07aa5f75399e0 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <B56489@freescale.com> +Date: Wed, 20 Jan 2016 15:52:25 +0800 +Subject: [PATCH 098/113] mtd: fsl-quadspi: Rename SEQID_QUAD_READ to + SEQID_READ + +There are some read modes for flash, such as NORMAL, FAST, +QUAD, DDR QUAD. These modes will use the identical lut table base +So rename SEQID_QUAD_READ to SEQID_READ. + +Signed-off-by: Yunhui Cui <B56489@freescale.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 8 ++++---- + 1 file changed, 4 insertions(+), 4 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -193,7 +193,7 @@ + #define QUADSPI_LUT_NUM 64 + + /* SEQID -- we can have 16 seqids at most. */ +-#define SEQID_QUAD_READ 0 ++#define SEQID_READ 0 + #define SEQID_WREN 1 + #define SEQID_WRDI 2 + #define SEQID_RDSR 3 +@@ -386,8 +386,8 @@ static void fsl_qspi_init_lut(struct fsl + for (i = 0; i < QUADSPI_LUT_NUM; i++) + qspi_writel(q, 0, base + QUADSPI_LUT_BASE + i * 4); + +- /* Quad Read */ +- lut_base = SEQID_QUAD_READ * 4; ++ /* Read */ ++ lut_base = SEQID_READ * 4; + + qspi_writel(q, LUT0(CMD, PAD1, read_op) | LUT1(ADDR, PAD1, addrlen), + base + QUADSPI_LUT(lut_base)); +@@ -468,7 +468,7 @@ static int fsl_qspi_get_seqid(struct fsl + { + switch (cmd) { + case SPINOR_OP_READ_1_1_4: +- return SEQID_QUAD_READ; ++ return SEQID_READ; + case SPINOR_OP_WREN: + return SEQID_WREN; + case SPINOR_OP_WRDI: diff --git a/target/linux/layerscape/patches-4.4/1099-mtd-spi-nor-fsl-quadspi-Add-fast-read-mode-support.patch b/target/linux/layerscape/patches-4.4/1099-mtd-spi-nor-fsl-quadspi-Add-fast-read-mode-support.patch new file mode 100644 index 0000000..ac99ba9 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1099-mtd-spi-nor-fsl-quadspi-Add-fast-read-mode-support.patch @@ -0,0 +1,72 @@ +From c501cdf57682265b72a8180c06e4a01dc2978375 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <B56489@freescale.com> +Date: Mon, 1 Feb 2016 18:26:23 +0800 +Subject: [PATCH 099/113] mtd:spi-nor:fsl-quadspi:Add fast-read mode support + +The qspi driver add generic fast-read mode for different +flash venders. There are some different board flash work on +different mode, such fast-read, quad-mode. +So we have to modify the third entrace parameter of spi_nor_scan(). + +Signed-off-by: Yunhui Cui <B56489@freescale.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 27 +++++++++++++++++++++------ + 1 file changed, 21 insertions(+), 6 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -389,11 +389,21 @@ static void fsl_qspi_init_lut(struct fsl + /* Read */ + lut_base = SEQID_READ * 4; + +- qspi_writel(q, LUT0(CMD, PAD1, read_op) | LUT1(ADDR, PAD1, addrlen), +- base + QUADSPI_LUT(lut_base)); +- qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | +- LUT1(FSL_READ, PAD4, rxfifo), +- base + QUADSPI_LUT(lut_base + 1)); ++ if (nor->flash_read == SPI_NOR_FAST) { ++ qspi_writel(q, LUT0(CMD, PAD1, read_op) | ++ LUT1(ADDR, PAD1, addrlen), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | ++ LUT1(FSL_READ, PAD1, rxfifo), ++ base + QUADSPI_LUT(lut_base + 1)); ++ } else if (nor->flash_read == SPI_NOR_QUAD) { ++ qspi_writel(q, LUT0(CMD, PAD1, read_op) | ++ LUT1(ADDR, PAD1, addrlen), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | ++ LUT1(FSL_READ, PAD4, rxfifo), ++ base + QUADSPI_LUT(lut_base + 1)); ++ } + + /* Write enable */ + lut_base = SEQID_WREN * 4; +@@ -468,6 +478,7 @@ static int fsl_qspi_get_seqid(struct fsl + { + switch (cmd) { + case SPINOR_OP_READ_1_1_4: ++ case SPINOR_OP_READ_FAST: + return SEQID_READ; + case SPINOR_OP_WREN: + return SEQID_WREN; +@@ -964,6 +975,7 @@ static int fsl_qspi_probe(struct platfor + struct spi_nor *nor; + struct mtd_info *mtd; + int ret, i = 0; ++ enum read_mode mode = SPI_NOR_QUAD; + + q = devm_kzalloc(dev, sizeof(*q), GFP_KERNEL); + if (!q) +@@ -1065,7 +1077,10 @@ static int fsl_qspi_probe(struct platfor + /* set the chip address for READID */ + fsl_qspi_set_base_addr(q, nor); + +- ret = spi_nor_scan(nor, NULL, SPI_NOR_QUAD); ++ ret = of_property_read_bool(np, "m25p,fast-read"); ++ mode = (ret) ? SPI_NOR_FAST : SPI_NOR_QUAD; ++ ++ ret = spi_nor_scan(nor, NULL, mode); + if (ret) + goto mutex_failed; + diff --git a/target/linux/layerscape/patches-4.4/1100-mtd-spi_nor-Disable-Micron-flash-HW-protection.patch b/target/linux/layerscape/patches-4.4/1100-mtd-spi_nor-Disable-Micron-flash-HW-protection.patch new file mode 100644 index 0000000..733b62c --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1100-mtd-spi_nor-Disable-Micron-flash-HW-protection.patch @@ -0,0 +1,41 @@ +From e892dea7229d56b75c46a76b9039f9e179584a91 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <B56489@freescale.com> +Date: Mon, 1 Feb 2016 18:48:49 +0800 +Subject: [PATCH 100/113] mtd:spi_nor: Disable Micron flash HW protection + +For Micron family ,The status register write enable/disable bit, +provides hardware data protection for the device. +When the enable/disable bit is set to 1, the status register +nonvolatile bits become read-only and the WRITE STATUS REGISTER +operation will not execute. + +Signed-off-by: Yunhui Cui <B56489@freescale.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 9 +++++++++ + 1 file changed, 9 insertions(+) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -39,6 +39,7 @@ + + #define SPI_NOR_MAX_ID_LEN 6 + #define SPI_NOR_MAX_ADDR_WIDTH 4 ++#define SPI_NOR_MICRON_WRITE_ENABLE 0x7f + + struct flash_info { + char *name; +@@ -1237,6 +1238,14 @@ int spi_nor_scan(struct spi_nor *nor, co + if (ret) + return ret; + ++ if (JEDEC_MFR(info) == SNOR_MFR_MICRON) { ++ ret = read_sr(nor); ++ ret &= SPI_NOR_MICRON_WRITE_ENABLE; ++ ++ write_enable(nor); ++ write_sr(nor, ret); ++ } ++ + if (!mtd->name) + mtd->name = dev_name(dev); + mtd->priv = nor; diff --git a/target/linux/layerscape/patches-4.4/1101-mtd-spi-nor-fsl-quadspi-extend-support-for-some-spec.patch b/target/linux/layerscape/patches-4.4/1101-mtd-spi-nor-fsl-quadspi-extend-support-for-some-spec.patch new file mode 100644 index 0000000..471ce42 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1101-mtd-spi-nor-fsl-quadspi-extend-support-for-some-spec.patch @@ -0,0 +1,122 @@ +From acfc6e9b34b3b3ca0d8bbe366dd08b0fac21c740 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Tue, 2 Feb 2016 12:21:12 +0800 +Subject: [PATCH 101/113] mtd: spi-nor: fsl-quadspi: extend support for some + special requerment. + +Add extra info in LUT table to support some special requerments. +Spansion S25FS-S family flash need some special operations. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 44 +++++++++++++++++++++++++++++++++++-- + include/linux/mtd/spi-nor.h | 4 ++++ + 2 files changed, 46 insertions(+), 2 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -205,6 +205,9 @@ + #define SEQID_RDCR 9 + #define SEQID_EN4B 10 + #define SEQID_BRWR 11 ++#define SEQID_RDAR 12 ++#define SEQID_WRAR 13 ++ + + #define QUADSPI_MIN_IOMAP SZ_4M + +@@ -470,6 +473,28 @@ static void fsl_qspi_init_lut(struct fsl + qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_BRWR), + base + QUADSPI_LUT(lut_base)); + ++ /* ++ * Read any device register. ++ * Used for Spansion S25FS-S family flash only. ++ */ ++ lut_base = SEQID_RDAR * 4; ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_SPANSION_RDAR) | ++ LUT1(ADDR, PAD1, ADDR24BIT), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(DUMMY, PAD1, 8) | LUT1(FSL_READ, PAD1, 1), ++ base + QUADSPI_LUT(lut_base + 1)); ++ ++ /* ++ * Write any device register. ++ * Used for Spansion S25FS-S family flash only. ++ */ ++ lut_base = SEQID_WRAR * 4; ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_SPANSION_WRAR) | ++ LUT1(ADDR, PAD1, ADDR24BIT), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(FSL_WRITE, PAD1, 1), ++ base + QUADSPI_LUT(lut_base + 1)); ++ + fsl_qspi_lock_lut(q); + } + +@@ -477,9 +502,15 @@ static void fsl_qspi_init_lut(struct fsl + static int fsl_qspi_get_seqid(struct fsl_qspi *q, u8 cmd) + { + switch (cmd) { ++ case SPINOR_OP_READ4_1_1_4: + case SPINOR_OP_READ_1_1_4: + case SPINOR_OP_READ_FAST: ++ case SPINOR_OP_READ4_FAST: + return SEQID_READ; ++ case SPINOR_OP_SPANSION_RDAR: ++ return SEQID_RDAR; ++ case SPINOR_OP_SPANSION_WRAR: ++ return SEQID_WRAR; + case SPINOR_OP_WREN: + return SEQID_WREN; + case SPINOR_OP_WRDI: +@@ -491,6 +522,7 @@ static int fsl_qspi_get_seqid(struct fsl + case SPINOR_OP_CHIP_ERASE: + return SEQID_CHIP_ERASE; + case SPINOR_OP_PP: ++ case SPINOR_OP_PP_4B: + return SEQID_PP; + case SPINOR_OP_RDID: + return SEQID_RDID; +@@ -830,8 +862,12 @@ static int fsl_qspi_read_reg(struct spi_ + { + int ret; + struct fsl_qspi *q = nor->priv; ++ u32 to = 0; ++ ++ if (opcode == SPINOR_OP_SPANSION_RDAR) ++ memcpy(&to, nor->cmd_buf, 4); + +- ret = fsl_qspi_runcmd(q, opcode, 0, len); ++ ret = fsl_qspi_runcmd(q, opcode, to, len); + if (ret) + return ret; + +@@ -843,9 +879,13 @@ static int fsl_qspi_write_reg(struct spi + { + struct fsl_qspi *q = nor->priv; + int ret; ++ u32 to = 0; ++ ++ if (opcode == SPINOR_OP_SPANSION_WRAR) ++ memcpy(&to, nor->cmd_buf, 4); + + if (!buf) { +- ret = fsl_qspi_runcmd(q, opcode, 0, 1); ++ ret = fsl_qspi_runcmd(q, opcode, to, 1); + if (ret) + return ret; + +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -74,6 +74,10 @@ + /* Used for Spansion flashes only. */ + #define SPINOR_OP_BRWR 0x17 /* Bank register write */ + ++/* Used for Spansion S25FS-S family flash only. */ ++#define SPINOR_OP_SPANSION_RDAR 0x65 /* Read any device register */ ++#define SPINOR_OP_SPANSION_WRAR 0x71 /* Write any device register */ ++ + /* Used for Micron flashes only. */ + #define SPINOR_OP_RD_EVCR 0x65 /* Read EVCR register */ + #define SPINOR_OP_WD_EVCR 0x61 /* Write EVCR register */ diff --git a/target/linux/layerscape/patches-4.4/1102-mtd-spi-nor-fsl-quadspi-Support-qspi-for-ls2080a.patch b/target/linux/layerscape/patches-4.4/1102-mtd-spi-nor-fsl-quadspi-Support-qspi-for-ls2080a.patch new file mode 100644 index 0000000..da9dd72 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1102-mtd-spi-nor-fsl-quadspi-Support-qspi-for-ls2080a.patch @@ -0,0 +1,83 @@ +From d2d88e3432d68b11b0add84bd15a3aadaf44f1c1 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <B56489@freescale.com> +Date: Mon, 28 Dec 2015 18:25:56 +0800 +Subject: [PATCH 102/113] mtd: spi-nor: fsl-quadspi:Support qspi for ls2080a + +There is a hardware feature that qspi_amba_base is added +internally by SOC design on ls2080a. So as to software, the driver +need support to the feature. + +Signed-off-by: Yunhui Cui <B56489@freescale.com> +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 24 ++++++++++++++++++++++-- + 1 file changed, 22 insertions(+), 2 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -41,6 +41,8 @@ + #define QUADSPI_QUIRK_TKT253890 (1 << 2) + /* Controller cannot wake up from wait mode, TKT245618 */ + #define QUADSPI_QUIRK_TKT245618 (1 << 3) ++/* QSPI_AMBA_BASE is internally added by SOC design */ ++#define QUADSPI_AMBA_BASE_INTERNAL (0x10000) + + /* The registers */ + #define QUADSPI_MCR 0x00 +@@ -217,6 +219,7 @@ enum fsl_qspi_devtype { + FSL_QUADSPI_IMX7D, + FSL_QUADSPI_IMX6UL, + FSL_QUADSPI_LS1021A, ++ FSL_QUADSPI_LS2080A, + }; + + struct fsl_qspi_devtype_data { +@@ -270,6 +273,14 @@ static struct fsl_qspi_devtype_data ls10 + .driver_data = 0, + }; + ++static struct fsl_qspi_devtype_data ls2080a_data = { ++ .devtype = FSL_QUADSPI_LS2080A, ++ .rxfifo = 128, ++ .txfifo = 64, ++ .ahb_buf_size = 1024, ++ .driver_data = QUADSPI_AMBA_BASE_INTERNAL, ++}; ++ + #define FSL_QSPI_MAX_CHIP 4 + struct fsl_qspi { + struct spi_nor nor[FSL_QSPI_MAX_CHIP]; +@@ -312,6 +323,11 @@ static inline int needs_wakeup_wait_mode + return q->devtype_data->driver_data & QUADSPI_QUIRK_TKT245618; + } + ++static inline int has_added_amba_base_internal(struct fsl_qspi *q) ++{ ++ return q->devtype_data->driver_data & QUADSPI_AMBA_BASE_INTERNAL; ++} ++ + /* + * R/W functions for big- or little-endian registers: + * The qSPI controller's endian is independent of the CPU core's endian. +@@ -558,8 +574,11 @@ fsl_qspi_runcmd(struct fsl_qspi *q, u8 c + /* save the reg */ + reg = qspi_readl(q, base + QUADSPI_MCR); + +- qspi_writel(q, q->memmap_phy + q->chip_base_addr + addr, +- base + QUADSPI_SFAR); ++ if (has_added_amba_base_internal(q)) ++ qspi_writel(q, q->chip_base_addr + addr, base + QUADSPI_SFAR); ++ else ++ qspi_writel(q, q->memmap_phy + q->chip_base_addr + addr, ++ base + QUADSPI_SFAR); + qspi_writel(q, QUADSPI_RBCT_WMRK_MASK | QUADSPI_RBCT_RXBRD_USEIPS, + base + QUADSPI_RBCT); + qspi_writel(q, reg | QUADSPI_MCR_CLR_RXF_MASK, base + QUADSPI_MCR); +@@ -849,6 +868,7 @@ static const struct of_device_id fsl_qsp + { .compatible = "fsl,imx7d-qspi", .data = (void *)&imx7d_data, }, + { .compatible = "fsl,imx6ul-qspi", .data = (void *)&imx6ul_data, }, + { .compatible = "fsl,ls1021a-qspi", .data = (void *)&ls1021a_data, }, ++ { .compatible = "fsl,ls2080a-qspi", .data = (void *)&ls2080a_data, }, + { /* sentinel */ } + }; + MODULE_DEVICE_TABLE(of, fsl_qspi_dt_ids); diff --git a/target/linux/layerscape/patches-4.4/1103-mtd-spi-nor-Support-R-W-for-S25FS-S-family-flash.patch b/target/linux/layerscape/patches-4.4/1103-mtd-spi-nor-Support-R-W-for-S25FS-S-family-flash.patch new file mode 100644 index 0000000..92041e9 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1103-mtd-spi-nor-Support-R-W-for-S25FS-S-family-flash.patch @@ -0,0 +1,110 @@ +From 0878404f549021e7fe0a49ae0454cf53fd452add Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Tue, 2 Feb 2016 12:00:27 +0800 +Subject: [PATCH 103/113] mtd: spi-nor: Support R/W for S25FS-S family flash + +With the physical sectors combination, S25FS-S family flash +requires some special operations for read/write functions. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 60 +++++++++++++++++++++++++++++++++++++++++ + 1 file changed, 60 insertions(+) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -40,6 +40,10 @@ + #define SPI_NOR_MAX_ID_LEN 6 + #define SPI_NOR_MAX_ADDR_WIDTH 4 + #define SPI_NOR_MICRON_WRITE_ENABLE 0x7f ++/* Added for S25FS-S family flash */ ++#define SPINOR_CONFIG_REG3_OFFSET 0x800004 ++#define CR3V_4KB_ERASE_UNABLE 0x8 ++#define SPINOR_S25FS_FAMILY_ID 0x81 + + struct flash_info { + char *name; +@@ -73,6 +77,8 @@ struct flash_info { + }; + + #define JEDEC_MFR(info) ((info)->id[0]) ++#define EXT_ID(info) ((info)->id[5]) ++ + + static const struct flash_info *spi_nor_match_id(const char *name); + +@@ -784,6 +790,7 @@ static const struct flash_info spi_nor_i + */ + { "s25sl032p", INFO(0x010215, 0x4d00, 64 * 1024, 64, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25sl064p", INFO(0x010216, 0x4d00, 64 * 1024, 128, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, ++ { "s25fs256s1", INFO6(0x010219, 0x4d0181, 64 * 1024, 512, 0)}, + { "s25fl256s0", INFO(0x010219, 0x4d00, 256 * 1024, 128, 0) }, + { "s25fl256s1", INFO(0x010219, 0x4d01, 64 * 1024, 512, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fl512s", INFO(0x010220, 0x4d00, 256 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, +@@ -908,6 +915,53 @@ static const struct flash_info *spi_nor_ + return ERR_PTR(-ENODEV); + } + ++/* ++ * The S25FS-S family physical sectors may be configured as a ++ * hybrid combination of eight 4-kB parameter sectors ++ * at the top or bottom of the address space with all ++ * but one of the remaining sectors being uniform size. ++ * The Parameter Sector Erase commands (20h or 21h) must ++ * be used to erase the 4-kB parameter sectors individually. ++ * The Sector (uniform sector) Erase commands (D8h or DCh) ++ * must be used to erase any of the remaining ++ * sectors, including the portion of highest or lowest address ++ * sector that is not overlaid by the parameter sectors. ++ * The uniform sector erase command has no effect on parameter sectors. ++ */ ++static int spansion_s25fs_disable_4kb_erase(struct spi_nor *nor) ++{ ++ struct fsl_qspi *q; ++ u32 cr3v_addr = SPINOR_CONFIG_REG3_OFFSET; ++ u8 cr3v = 0x0; ++ int ret = 0x0; ++ ++ q = nor->priv; ++ ++ nor->cmd_buf[2] = cr3v_addr >> 16; ++ nor->cmd_buf[1] = cr3v_addr >> 8; ++ nor->cmd_buf[0] = cr3v_addr >> 0; ++ ++ ret = nor->read_reg(nor, SPINOR_OP_SPANSION_RDAR, &cr3v, 1); ++ if (ret) ++ return ret; ++ if (cr3v & CR3V_4KB_ERASE_UNABLE) ++ return 0; ++ ret = nor->write_reg(nor, SPINOR_OP_WREN, NULL, 0); ++ if (ret) ++ return ret; ++ cr3v = CR3V_4KB_ERASE_UNABLE; ++ nor->program_opcode = SPINOR_OP_SPANSION_WRAR; ++ nor->write(nor, cr3v_addr, 1, &cr3v); ++ ++ ret = nor->read_reg(nor, SPINOR_OP_SPANSION_RDAR, &cr3v, 1); ++ if (ret) ++ return ret; ++ if (!(cr3v & CR3V_4KB_ERASE_UNABLE)) ++ return -EPERM; ++ ++ return 0; ++} ++ + static int spi_nor_read(struct mtd_info *mtd, loff_t from, size_t len, + size_t *retlen, u_char *buf) + { +@@ -1246,6 +1300,12 @@ int spi_nor_scan(struct spi_nor *nor, co + write_sr(nor, ret); + } + ++ if (EXT_ID(info) == SPINOR_S25FS_FAMILY_ID) { ++ ret = spansion_s25fs_disable_4kb_erase(nor); ++ if (ret) ++ return ret; ++ } ++ + if (!mtd->name) + mtd->name = dev_name(dev); + mtd->priv = nor; diff --git a/target/linux/layerscape/patches-4.4/1104-mtd-fsl-quadspi-Add-quad-mode-for-flash-n25q128.patch b/target/linux/layerscape/patches-4.4/1104-mtd-fsl-quadspi-Add-quad-mode-for-flash-n25q128.patch new file mode 100644 index 0000000..daf34e3 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1104-mtd-fsl-quadspi-Add-quad-mode-for-flash-n25q128.patch @@ -0,0 +1,112 @@ +From 23cd071c47c064d56921975d196dc22177069dea Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Wed, 24 Feb 2016 15:14:01 +0800 +Subject: [PATCH 104/113] mtd: fsl-quadspi: Add quad mode for flash n25q128 + +Add some lut_tables to support quad mode for flash n25q128 +on the board ls1021a-twr and solve flash Spansion and Micron +command conflict. +In switch {}, The value of command SPINOR_OP_RD_EVCR and +SPINOR_OP_SPANSION_RDAR is the same. They have to share +the same seq_id: SEQID_RDAR_OR_RD_EVCR. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 47 ++++++++++++++++++++++++++++--------- + 1 file changed, 36 insertions(+), 11 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -207,9 +207,9 @@ + #define SEQID_RDCR 9 + #define SEQID_EN4B 10 + #define SEQID_BRWR 11 +-#define SEQID_RDAR 12 ++#define SEQID_RDAR_OR_RD_EVCR 12 + #define SEQID_WRAR 13 +- ++#define SEQID_WD_EVCR 14 + + #define QUADSPI_MIN_IOMAP SZ_4M + +@@ -393,6 +393,7 @@ static void fsl_qspi_init_lut(struct fsl + int rxfifo = q->devtype_data->rxfifo; + u32 lut_base; + int i; ++ const struct fsl_qspi_devtype_data *devtype_data = q->devtype_data; + + struct spi_nor *nor = &q->nor[0]; + u8 addrlen = (nor->addr_width == 3) ? ADDR24BIT : ADDR32BIT; +@@ -489,16 +490,26 @@ static void fsl_qspi_init_lut(struct fsl + qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_BRWR), + base + QUADSPI_LUT(lut_base)); + ++ + /* +- * Read any device register. +- * Used for Spansion S25FS-S family flash only. ++ * Flash Micron and Spansion command confilict ++ * use the same value 0x65. But it indicates different meaning. + */ +- lut_base = SEQID_RDAR * 4; +- qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_SPANSION_RDAR) | +- LUT1(ADDR, PAD1, ADDR24BIT), +- base + QUADSPI_LUT(lut_base)); +- qspi_writel(q, LUT0(DUMMY, PAD1, 8) | LUT1(FSL_READ, PAD1, 1), +- base + QUADSPI_LUT(lut_base + 1)); ++ lut_base = SEQID_RDAR_OR_RD_EVCR * 4; ++ if (devtype_data->devtype == FSL_QUADSPI_LS2080A) { ++ /* ++ * Read any device register. ++ * Used for Spansion S25FS-S family flash only. ++ */ ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_SPANSION_RDAR) | ++ LUT1(ADDR, PAD1, ADDR24BIT), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(DUMMY, PAD1, 8) | LUT1(FSL_READ, PAD1, 1), ++ base + QUADSPI_LUT(lut_base + 1)); ++ } else { ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_RD_EVCR), ++ base + QUADSPI_LUT(lut_base)); ++ } + + /* + * Write any device register. +@@ -511,6 +522,11 @@ static void fsl_qspi_init_lut(struct fsl + qspi_writel(q, LUT0(FSL_WRITE, PAD1, 1), + base + QUADSPI_LUT(lut_base + 1)); + ++ /* Write EVCR register */ ++ lut_base = SEQID_WD_EVCR * 4; ++ qspi_writel(q, LUT0(CMD, PAD1, SPINOR_OP_WD_EVCR), ++ base + QUADSPI_LUT(lut_base)); ++ + fsl_qspi_lock_lut(q); + } + +@@ -523,8 +539,15 @@ static int fsl_qspi_get_seqid(struct fsl + case SPINOR_OP_READ_FAST: + case SPINOR_OP_READ4_FAST: + return SEQID_READ; ++ /* ++ * Spansion & Micron use the same command value 0x65 ++ * Spansion: SPINOR_OP_SPANSION_RDAR, read any register. ++ * Micron: SPINOR_OP_RD_EVCR, ++ * read enhanced volatile configuration register. ++ * case SPINOR_OP_RD_EVCR: ++ */ + case SPINOR_OP_SPANSION_RDAR: +- return SEQID_RDAR; ++ return SEQID_RDAR_OR_RD_EVCR; + case SPINOR_OP_SPANSION_WRAR: + return SEQID_WRAR; + case SPINOR_OP_WREN: +@@ -550,6 +573,8 @@ static int fsl_qspi_get_seqid(struct fsl + return SEQID_EN4B; + case SPINOR_OP_BRWR: + return SEQID_BRWR; ++ case SPINOR_OP_WD_EVCR: ++ return SEQID_WD_EVCR; + default: + if (cmd == q->nor[0].erase_opcode) + return SEQID_SE; diff --git a/target/linux/layerscape/patches-4.4/1105-mtd-spi-nor-add-DDR-quad-read-support.patch b/target/linux/layerscape/patches-4.4/1105-mtd-spi-nor-add-DDR-quad-read-support.patch new file mode 100644 index 0000000..bed44b8 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1105-mtd-spi-nor-add-DDR-quad-read-support.patch @@ -0,0 +1,181 @@ +From 924f021c0344554a4b61746e5c4dcfc91d618ce2 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Thu, 18 Feb 2016 16:41:53 +0800 +Subject: [PATCH 105/113] mtd: spi-nor: add DDR quad read support + +This patch adds the DDR quad read support by the following: + + [1] add SPI_NOR_DDR_QUAD read mode. + + [2] add DDR Quad read opcodes: + SPINOR_OP_READ_1_4_4_D / SPINOR_OP_READ4_1_4_4_D + + [3] add set_ddr_quad_mode() to initialize for the DDR quad read. + Currently it only works for Spansion NOR. + + [4] set dummy with 6 for Spansion family +Test this patch for Spansion s25fl128s NOR flash. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 53 ++++++++++++++++++++++++++++++++++++----- + include/linux/mtd/spi-nor.h | 8 +++++-- + 2 files changed, 53 insertions(+), 8 deletions(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -73,7 +73,8 @@ struct flash_info { + #define SECT_4K_PMC 0x10 /* SPINOR_OP_BE_4K_PMC works uniformly */ + #define SPI_NOR_DUAL_READ 0x20 /* Flash supports Dual Read */ + #define SPI_NOR_QUAD_READ 0x40 /* Flash supports Quad Read */ +-#define USE_FSR 0x80 /* use flag status register */ ++#define SPI_NOR_DDR_QUAD_READ 0x80 /* Flash supports DDR Quad Read */ ++#define USE_FSR 0x100 /* use flag status register */ + }; + + #define JEDEC_MFR(info) ((info)->id[0]) +@@ -144,13 +145,17 @@ static int read_cr(struct spi_nor *nor) + * It can be used to support more commands with + * different dummy cycle requirements. + */ +-static inline int spi_nor_read_dummy_cycles(struct spi_nor *nor) ++static inline int spi_nor_read_dummy_cycles(struct spi_nor *nor, ++ const struct flash_info *info) + { + switch (nor->flash_read) { + case SPI_NOR_FAST: + case SPI_NOR_DUAL: + case SPI_NOR_QUAD: + return 8; ++ case SPI_NOR_DDR_QUAD: ++ if (JEDEC_MFR(info) == SNOR_MFR_SPANSION) ++ return 6; + case SPI_NOR_NORMAL: + return 0; + } +@@ -797,7 +802,8 @@ static const struct flash_info spi_nor_i + { "s70fl01gs", INFO(0x010221, 0x4d00, 256 * 1024, 256, 0) }, + { "s25sl12800", INFO(0x012018, 0x0300, 256 * 1024, 64, 0) }, + { "s25sl12801", INFO(0x012018, 0x0301, 64 * 1024, 256, 0) }, +- { "s25fl128s", INFO6(0x012018, 0x4d0180, 64 * 1024, 256, SECT_4K | SPI_NOR_QUAD_READ) }, ++ { "s25fl128s", INFO6(0x012018, 0x4d0180, 64 * 1024, 256, SECT_4K | SPI_NOR_QUAD_READ ++ | SPI_NOR_DDR_QUAD_READ) }, + { "s25fl129p0", INFO(0x012018, 0x4d00, 256 * 1024, 64, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fl129p1", INFO(0x012018, 0x4d01, 64 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25sl004a", INFO(0x010212, 0, 64 * 1024, 8, 0) }, +@@ -1186,6 +1192,23 @@ static int spansion_quad_enable(struct s + return 0; + } + ++static int set_ddr_quad_mode(struct spi_nor *nor, const struct flash_info *info) ++{ ++ int status; ++ ++ switch (JEDEC_MFR(info)) { ++ case SNOR_MFR_SPANSION: ++ status = spansion_quad_enable(nor); ++ if (status) { ++ dev_err(nor->dev, "Spansion DDR quad-read not enabled\n"); ++ return status; ++ } ++ return status; ++ default: ++ return -EINVAL; ++ } ++} ++ + static int set_quad_mode(struct spi_nor *nor, const struct flash_info *info) + { + int status; +@@ -1375,8 +1398,15 @@ int spi_nor_scan(struct spi_nor *nor, co + if (info->flags & SPI_NOR_NO_FR) + nor->flash_read = SPI_NOR_NORMAL; + +- /* Quad/Dual-read mode takes precedence over fast/normal */ +- if (mode == SPI_NOR_QUAD && info->flags & SPI_NOR_QUAD_READ) { ++ /* DDR Quad/Quad/Dual-read mode takes precedence over fast/normal */ ++ if (mode == SPI_NOR_DDR_QUAD && info->flags & SPI_NOR_DDR_QUAD_READ) { ++ ret = set_ddr_quad_mode(nor, info); ++ if (ret) { ++ dev_err(dev, "DDR quad mode not supported\n"); ++ return ret; ++ } ++ nor->flash_read = SPI_NOR_DDR_QUAD; ++ } else if (mode == SPI_NOR_QUAD && info->flags & SPI_NOR_QUAD_READ) { + ret = set_quad_mode(nor, info); + if (ret) { + dev_err(dev, "quad mode not supported\n"); +@@ -1389,6 +1419,14 @@ int spi_nor_scan(struct spi_nor *nor, co + + /* Default commands */ + switch (nor->flash_read) { ++ case SPI_NOR_DDR_QUAD: ++ if (JEDEC_MFR(info) == SNOR_MFR_SPANSION) { /* Spansion */ ++ nor->read_opcode = SPINOR_OP_READ_1_4_4_D; ++ } else { ++ dev_err(dev, "DDR Quad Read is not supported.\n"); ++ return -EINVAL; ++ } ++ break; + case SPI_NOR_QUAD: + nor->read_opcode = SPINOR_OP_READ_1_1_4; + break; +@@ -1416,6 +1454,9 @@ int spi_nor_scan(struct spi_nor *nor, co + if (JEDEC_MFR(info) == SNOR_MFR_SPANSION) { + /* Dedicated 4-byte command set */ + switch (nor->flash_read) { ++ case SPI_NOR_DDR_QUAD: ++ nor->read_opcode = SPINOR_OP_READ4_1_4_4_D; ++ break; + case SPI_NOR_QUAD: + nor->read_opcode = SPINOR_OP_READ4_1_1_4; + break; +@@ -1445,7 +1486,7 @@ int spi_nor_scan(struct spi_nor *nor, co + return -EINVAL; + } + +- nor->read_dummy = spi_nor_read_dummy_cycles(nor); ++ nor->read_dummy = spi_nor_read_dummy_cycles(nor, info); + + dev_info(dev, "%s (%lld Kbytes)\n", info->name, + (long long)mtd->size >> 10); +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -30,10 +30,11 @@ + + /* + * Note on opcode nomenclature: some opcodes have a format like +- * SPINOR_OP_FUNCTION{4,}_x_y_z. The numbers x, y, and z stand for the number ++ * SPINOR_OP_FUNCTION{4,}_x_y_z{_D}. The numbers x, y,and z stand for the number + * of I/O lines used for the opcode, address, and data (respectively). The + * FUNCTION has an optional suffix of '4', to represent an opcode which +- * requires a 4-byte (32-bit) address. ++ * requires a 4-byte (32-bit) address. The suffix of 'D' stands for the ++ * DDR mode. + */ + + /* Flash opcodes. */ +@@ -44,6 +45,7 @@ + #define SPINOR_OP_READ_FAST 0x0b /* Read data bytes (high frequency) */ + #define SPINOR_OP_READ_1_1_2 0x3b /* Read data bytes (Dual SPI) */ + #define SPINOR_OP_READ_1_1_4 0x6b /* Read data bytes (Quad SPI) */ ++#define SPINOR_OP_READ_1_4_4_D 0xed /* Read data bytes (DDR Quad SPI) */ + #define SPINOR_OP_PP 0x02 /* Page program (up to 256 bytes) */ + #define SPINOR_OP_BE_4K 0x20 /* Erase 4KiB block */ + #define SPINOR_OP_BE_4K_PMC 0xd7 /* Erase 4KiB block on PMC chips */ +@@ -59,6 +61,7 @@ + #define SPINOR_OP_READ4_FAST 0x0c /* Read data bytes (high frequency) */ + #define SPINOR_OP_READ4_1_1_2 0x3c /* Read data bytes (Dual SPI) */ + #define SPINOR_OP_READ4_1_1_4 0x6c /* Read data bytes (Quad SPI) */ ++#define SPINOR_OP_READ4_1_4_4_D 0xee /* Read data bytes (DDR Quad SPI) */ + #define SPINOR_OP_PP_4B 0x12 /* Page program (up to 256 bytes) */ + #define SPINOR_OP_SE_4B 0xdc /* Sector erase (usually 64KiB) */ + +@@ -107,6 +110,7 @@ enum read_mode { + SPI_NOR_FAST, + SPI_NOR_DUAL, + SPI_NOR_QUAD, ++ SPI_NOR_DDR_QUAD, + }; + + #define SPI_NOR_MAX_CMD_SIZE 8 diff --git a/target/linux/layerscape/patches-4.4/1106-mtd-fsl-quadspi-add-DDR-quad-read-for-Spansion.patch b/target/linux/layerscape/patches-4.4/1106-mtd-fsl-quadspi-add-DDR-quad-read-for-Spansion.patch new file mode 100644 index 0000000..3a24260 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1106-mtd-fsl-quadspi-add-DDR-quad-read-for-Spansion.patch @@ -0,0 +1,122 @@ +From 16eb35ceea5b43e6f64c1a869721ea86c0da5260 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Thu, 25 Feb 2016 10:19:15 +0800 +Subject: [PATCH 106/113] mtd: fsl-quadspi: add DDR quad read for Spansion + +Add the DDR quad read support for the fsl-quadspi driver. +And, add the Spansion s25fl128s NOR flash ddr quad mode +support. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 57 +++++++++++++++++++++++++++++++++++++ + 1 file changed, 57 insertions(+) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -296,6 +296,7 @@ struct fsl_qspi { + u32 nor_size; + u32 nor_num; + u32 clk_rate; ++ u32 ddr_smp; + unsigned int chip_base_addr; /* We may support two chips. */ + bool has_second_chip; + bool big_endian; +@@ -423,6 +424,19 @@ static void fsl_qspi_init_lut(struct fsl + qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | + LUT1(FSL_READ, PAD4, rxfifo), + base + QUADSPI_LUT(lut_base + 1)); ++ } else if (nor->flash_read == SPI_NOR_DDR_QUAD) { ++ /* read mode : 1-4-4, such as Spansion s25fl128s. */ ++ qspi_writel(q, LUT0(CMD, PAD1, read_op) ++ | LUT1(ADDR_DDR, PAD4, addrlen), ++ base + QUADSPI_LUT(lut_base)); ++ ++ qspi_writel(q, LUT0(MODE_DDR, PAD4, 0xff) ++ | LUT1(DUMMY, PAD1, read_dm), ++ base + QUADSPI_LUT(lut_base + 1)); ++ ++ qspi_writel(q, LUT0(FSL_READ_DDR, PAD4, rxfifo) ++ | LUT1(JMP_ON_CS, PAD1, 0), ++ base + QUADSPI_LUT(lut_base + 2)); + } + + /* Write enable */ +@@ -534,6 +548,8 @@ static void fsl_qspi_init_lut(struct fsl + static int fsl_qspi_get_seqid(struct fsl_qspi *q, u8 cmd) + { + switch (cmd) { ++ case SPINOR_OP_READ_1_4_4_D: ++ case SPINOR_OP_READ4_1_4_4_D: + case SPINOR_OP_READ4_1_1_4: + case SPINOR_OP_READ_1_1_4: + case SPINOR_OP_READ_FAST: +@@ -736,6 +752,32 @@ static void fsl_qspi_set_map_addr(struct + } + + /* ++ * enable controller ddr quad mode to support different ++ * vender flashes ddr quad mode. ++ */ ++static void set_ddr_quad_mode(struct fsl_qspi *q) ++{ ++ u32 reg, reg2; ++ ++ reg = qspi_readl(q, q->iobase + QUADSPI_MCR); ++ ++ /* Firstly, disable the module */ ++ qspi_writel(q, reg | QUADSPI_MCR_MDIS_MASK, q->iobase + QUADSPI_MCR); ++ ++ /* Set the Sampling Register for DDR */ ++ reg2 = qspi_readl(q, q->iobase + QUADSPI_SMPR); ++ reg2 &= ~QUADSPI_SMPR_DDRSMP_MASK; ++ reg2 |= (((q->ddr_smp) << QUADSPI_SMPR_DDRSMP_SHIFT) & ++ QUADSPI_SMPR_DDRSMP_MASK); ++ qspi_writel(q, reg2, q->iobase + QUADSPI_SMPR); ++ ++ /* Enable the module again (enable the DDR too) */ ++ reg |= QUADSPI_MCR_DDR_EN_MASK; ++ qspi_writel(q, reg, q->iobase + QUADSPI_MCR); ++ ++} ++ ++/* + * There are two different ways to read out the data from the flash: + * the "IP Command Read" and the "AHB Command Read". + * +@@ -775,6 +817,11 @@ static void fsl_qspi_init_abh_read(struc + seqid = fsl_qspi_get_seqid(q, q->nor[0].read_opcode); + qspi_writel(q, seqid << QUADSPI_BFGENCR_SEQID_SHIFT, + q->iobase + QUADSPI_BFGENCR); ++ ++ /* enable the DDR quad read */ ++ if (q->nor->flash_read == SPI_NOR_DDR_QUAD) ++ set_ddr_quad_mode(q); ++ + } + + /* This function was used to prepare and enable QSPI clock */ +@@ -1108,6 +1155,12 @@ static int fsl_qspi_probe(struct platfor + goto clk_failed; + } + ++ /* find ddrsmp value */ ++ ret = of_property_read_u32(dev->of_node, "fsl,ddr-sampling-point", ++ &q->ddr_smp); ++ if (ret) ++ q->ddr_smp = 0; ++ + /* find the irq */ + ret = platform_get_irq(pdev, 0); + if (ret < 0) { +@@ -1164,6 +1217,10 @@ static int fsl_qspi_probe(struct platfor + + ret = of_property_read_bool(np, "m25p,fast-read"); + mode = (ret) ? SPI_NOR_FAST : SPI_NOR_QUAD; ++ /* Can we enable the DDR Quad Read? */ ++ ret = of_property_read_bool(np, "ddr-quad-read"); ++ if (ret) ++ mode = SPI_NOR_DDR_QUAD; + + ret = spi_nor_scan(nor, NULL, mode); + if (ret) diff --git a/target/linux/layerscape/patches-4.4/1107-mtd-fsl-quadspi-disable-AHB-buffer-prefetch.patch b/target/linux/layerscape/patches-4.4/1107-mtd-fsl-quadspi-disable-AHB-buffer-prefetch.patch new file mode 100644 index 0000000..19dff6c --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1107-mtd-fsl-quadspi-disable-AHB-buffer-prefetch.patch @@ -0,0 +1,67 @@ +From 50aac689d5be0a086f076cd4bc8b14ee0b9ab995 Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Tue, 8 Mar 2016 14:38:52 +0800 +Subject: [PATCH 107/113] mtd: fsl-quadspi: disable AHB buffer prefetch + +A-009282: QuadSPI: QuadSPI data pre-fetch can result in incorrect data +Affects: QuadSPI +Description: With AHB buffer prefetch enabled, the QuadSPI may return +incorrect data on the AHB +interface. The buffer pre-fetch is enabled if the fetch size as +configured either in the LUT or in +the BUFxCR register is greater than 8 bytes. +Impact: Only 64 bit read allowed. +Workaround: Keep the read data size to 64 bits (8 Bytes), which disables +the prefetch on the AHB buffer, +and prevents this issue from occurring. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 29 +++++++++++++++++++++++------ + 1 file changed, 23 insertions(+), 6 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -794,19 +794,36 @@ static void fsl_qspi_init_abh_read(struc + { + void __iomem *base = q->iobase; + int seqid; ++ const struct fsl_qspi_devtype_data *devtype_data = q->devtype_data; + + /* AHB configuration for access buffer 0/1/2 .*/ + qspi_writel(q, QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF0CR); + qspi_writel(q, QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF1CR); + qspi_writel(q, QUADSPI_BUFXCR_INVALID_MSTRID, base + QUADSPI_BUF2CR); ++ + /* +- * Set ADATSZ with the maximum AHB buffer size to improve the +- * read performance. ++ * Errata: A-009282: QuadSPI data prefetch may result in incorrect data ++ * Workaround: Keep the read data size to 64 bits (8 bytes). ++ * This disables the prefetch on the AHB buffer and ++ * prevents this issue from occurring. + */ +- qspi_writel(q, QUADSPI_BUF3CR_ALLMST_MASK | +- ((q->devtype_data->ahb_buf_size / 8) +- << QUADSPI_BUF3CR_ADATSZ_SHIFT), +- base + QUADSPI_BUF3CR); ++ if (devtype_data->devtype == FSL_QUADSPI_LS2080A || ++ devtype_data->devtype == FSL_QUADSPI_LS1021A) { ++ ++ qspi_writel(q, QUADSPI_BUF3CR_ALLMST_MASK | ++ (1 << QUADSPI_BUF3CR_ADATSZ_SHIFT), ++ base + QUADSPI_BUF3CR); ++ ++ } else { ++ /* ++ * Set ADATSZ with the maximum AHB buffer size to improve the ++ * read performance. ++ */ ++ qspi_writel(q, QUADSPI_BUF3CR_ALLMST_MASK | ++ ((q->devtype_data->ahb_buf_size / 8) ++ << QUADSPI_BUF3CR_ADATSZ_SHIFT), ++ base + QUADSPI_BUF3CR); ++ } + + /* We only use the buffer3 */ + qspi_writel(q, 0, base + QUADSPI_BUF0IND); diff --git a/target/linux/layerscape/patches-4.4/1108-mtd-fsl-quadspi-add-multi-flash-chip-R-W-on-ls2080a.patch b/target/linux/layerscape/patches-4.4/1108-mtd-fsl-quadspi-add-multi-flash-chip-R-W-on-ls2080a.patch new file mode 100644 index 0000000..4a7a5cb --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1108-mtd-fsl-quadspi-add-multi-flash-chip-R-W-on-ls2080a.patch @@ -0,0 +1,42 @@ +From d3a8ee41170ff9e5298ff354c77ff99439dfe2bf Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Thu, 10 Mar 2016 11:33:40 +0800 +Subject: [PATCH 108/113] mtd: fsl-quadspi: add multi flash chip R/W on + ls2080a + +There is a hardware feature that qspi_amba_base is added +internally by SOC design on ls2080a. so memmap_phy need not +be added in driver. If memmap_phy is added, the flash A1 +addr space is [0, memmap_phy] which far more than flash size. +The AMBA memory will be divided into four parts and assign to +every chipselect. Every channel will has two valid chipselects. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 14 ++++++++++---- + 1 file changed, 10 insertions(+), 4 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -744,11 +744,17 @@ static void fsl_qspi_set_map_addr(struct + { + int nor_size = q->nor_size; + void __iomem *base = q->iobase; ++ u32 mem_base; + +- qspi_writel(q, nor_size + q->memmap_phy, base + QUADSPI_SFA1AD); +- qspi_writel(q, nor_size * 2 + q->memmap_phy, base + QUADSPI_SFA2AD); +- qspi_writel(q, nor_size * 3 + q->memmap_phy, base + QUADSPI_SFB1AD); +- qspi_writel(q, nor_size * 4 + q->memmap_phy, base + QUADSPI_SFB2AD); ++ if (has_added_amba_base_internal(q)) ++ mem_base = 0x0; ++ else ++ mem_base = q->memmap_phy; ++ ++ qspi_writel(q, nor_size + mem_base, base + QUADSPI_SFA1AD); ++ qspi_writel(q, nor_size * 2 + mem_base, base + QUADSPI_SFA2AD); ++ qspi_writel(q, nor_size * 3 + mem_base, base + QUADSPI_SFB1AD); ++ qspi_writel(q, nor_size * 4 + mem_base, base + QUADSPI_SFB2AD); + } + + /* diff --git a/target/linux/layerscape/patches-4.4/1109-drivers-mtd-spi-nor-Enable-QSPI-Flash-in-Kernel.patch b/target/linux/layerscape/patches-4.4/1109-drivers-mtd-spi-nor-Enable-QSPI-Flash-in-Kernel.patch new file mode 100644 index 0000000..ec6423d --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1109-drivers-mtd-spi-nor-Enable-QSPI-Flash-in-Kernel.patch @@ -0,0 +1,36 @@ +From 70516f60de441829e7813c0b26567c8bda39c011 Mon Sep 17 00:00:00 2001 +From: Pratiyush Mohan Srivastava <pratiyush.srivastava@nxp.com> +Date: Sun, 24 Apr 2016 23:20:26 +0530 +Subject: [PATCH 109/113] drivers: mtd: spi-nor: Enable QSPI Flash in Kernel + +Enable read from QSPI flash, Write onto QSPI Flash and +erase QSPI Flash in Fast mode in Kernel. + +Signed-off-by: Pratiyush Mohan Srivastava <pratiyush.srivastava@nxp.com> +Signed-off-by: Prabhakar Kushwaha <prabhakar.kushwaha@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 3 +++ + 1 file changed, 3 insertions(+) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -796,6 +796,7 @@ static const struct flash_info spi_nor_i + { "s25sl032p", INFO(0x010215, 0x4d00, 64 * 1024, 64, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25sl064p", INFO(0x010216, 0x4d00, 64 * 1024, 128, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fs256s1", INFO6(0x010219, 0x4d0181, 64 * 1024, 512, 0)}, ++ { "s25fs512s", INFO6(0x010220, 0x4d0081, 128 * 1024, 512, 0)}, + { "s25fl256s0", INFO(0x010219, 0x4d00, 256 * 1024, 128, 0) }, + { "s25fl256s1", INFO(0x010219, 0x4d01, 64 * 1024, 512, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fl512s", INFO(0x010220, 0x4d00, 256 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, +@@ -962,9 +963,11 @@ static int spansion_s25fs_disable_4kb_er + ret = nor->read_reg(nor, SPINOR_OP_SPANSION_RDAR, &cr3v, 1); + if (ret) + return ret; ++/* + if (!(cr3v & CR3V_4KB_ERASE_UNABLE)) + return -EPERM; + ++*/ + return 0; + } + diff --git a/target/linux/layerscape/patches-4.4/1110-mtd-spi-nor-fsl-quad-add-flash-S25FS-extra-support.patch b/target/linux/layerscape/patches-4.4/1110-mtd-spi-nor-fsl-quad-add-flash-S25FS-extra-support.patch new file mode 100644 index 0000000..2b21b11 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1110-mtd-spi-nor-fsl-quad-add-flash-S25FS-extra-support.patch @@ -0,0 +1,157 @@ +From 034dd6241b55ab2256eecb845e941fa9b45da38e Mon Sep 17 00:00:00 2001 +From: Yunhui Cui <yunhui.cui@nxp.com> +Date: Thu, 28 Apr 2016 17:03:57 +0800 +Subject: [PATCH 110/113] mtd: spi-nor: fsl-quad: add flash S25FS extra + support + +[context adjustment] +not apply changes of arch/arm64/boot/dts/freescale/fsl-ls1012a-qds.dts + +There are some boards have the same QSPI controller but have +different vendor flash, So, the controller can use the same +compatible and share the driver, just for a different flash to do +the appropriate adaptation. Based on this, we need add the vendor +field in spi-nor, Because we will use the field to distribute +corresponding LUT for different flash operations. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +Signed-off-by: Pratiyush Mohan Srivastava <pratiyush.srivastava@nxp.com> +Signed-off-by: Prabhakar Kushwaha <prabhakar.kushwaha@nxp.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 47 ++++++++++++++++++++++++++++++------- + drivers/mtd/spi-nor/spi-nor.c | 5 ++-- + include/linux/mtd/spi-nor.h | 1 + + 3 files changed, 42 insertions(+), 11 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -213,6 +213,9 @@ + + #define QUADSPI_MIN_IOMAP SZ_4M + ++#define FLASH_VENDOR_SPANSION_FS "s25fs" ++#define SPANSION_S25FS_FAMILY (1 << 1) ++ + enum fsl_qspi_devtype { + FSL_QUADSPI_VYBRID, + FSL_QUADSPI_IMX6SX, +@@ -329,6 +332,18 @@ static inline int has_added_amba_base_in + return q->devtype_data->driver_data & QUADSPI_AMBA_BASE_INTERNAL; + } + ++static u32 fsl_get_nor_vendor(struct spi_nor *nor) ++{ ++ u32 vendor_id; ++ ++ if (nor->vendor) { ++ if (memcmp(nor->vendor, FLASH_VENDOR_SPANSION_FS, ++ sizeof(FLASH_VENDOR_SPANSION_FS) - 1)) ++ vendor_id = SPANSION_S25FS_FAMILY; ++ } ++ return vendor_id; ++} ++ + /* + * R/W functions for big- or little-endian registers: + * The qSPI controller's endian is independent of the CPU core's endian. +@@ -394,13 +409,15 @@ static void fsl_qspi_init_lut(struct fsl + int rxfifo = q->devtype_data->rxfifo; + u32 lut_base; + int i; +- const struct fsl_qspi_devtype_data *devtype_data = q->devtype_data; ++ u32 vendor; + + struct spi_nor *nor = &q->nor[0]; + u8 addrlen = (nor->addr_width == 3) ? ADDR24BIT : ADDR32BIT; + u8 read_op = nor->read_opcode; + u8 read_dm = nor->read_dummy; + ++ vendor = fsl_get_nor_vendor(nor); ++ + fsl_qspi_unlock_lut(q); + + /* Clear all the LUT table */ +@@ -418,12 +435,25 @@ static void fsl_qspi_init_lut(struct fsl + LUT1(FSL_READ, PAD1, rxfifo), + base + QUADSPI_LUT(lut_base + 1)); + } else if (nor->flash_read == SPI_NOR_QUAD) { +- qspi_writel(q, LUT0(CMD, PAD1, read_op) | +- LUT1(ADDR, PAD1, addrlen), +- base + QUADSPI_LUT(lut_base)); +- qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | +- LUT1(FSL_READ, PAD4, rxfifo), +- base + QUADSPI_LUT(lut_base + 1)); ++ if (q->nor_size == 0x4000000) { ++ read_op = 0xEC; ++ qspi_writel(q, ++ LUT0(CMD, PAD1, read_op) | LUT1(ADDR, PAD4, addrlen), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, ++ LUT0(MODE, PAD4, 0xff) | LUT1(DUMMY, PAD4, read_dm), ++ base + QUADSPI_LUT(lut_base + 1)); ++ qspi_writel(q, ++ LUT0(FSL_READ, PAD4, rxfifo), ++ base + QUADSPI_LUT(lut_base + 2)); ++ } else { ++ qspi_writel(q, LUT0(CMD, PAD1, read_op) | ++ LUT1(ADDR, PAD1, addrlen), ++ base + QUADSPI_LUT(lut_base)); ++ qspi_writel(q, LUT0(DUMMY, PAD1, read_dm) | ++ LUT1(FSL_READ, PAD4, rxfifo), ++ base + QUADSPI_LUT(lut_base + 1)); ++ } + } else if (nor->flash_read == SPI_NOR_DDR_QUAD) { + /* read mode : 1-4-4, such as Spansion s25fl128s. */ + qspi_writel(q, LUT0(CMD, PAD1, read_op) +@@ -510,7 +540,8 @@ static void fsl_qspi_init_lut(struct fsl + * use the same value 0x65. But it indicates different meaning. + */ + lut_base = SEQID_RDAR_OR_RD_EVCR * 4; +- if (devtype_data->devtype == FSL_QUADSPI_LS2080A) { ++ ++ if (vendor == SPANSION_S25FS_FAMILY) { + /* + * Read any device register. + * Used for Spansion S25FS-S family flash only. +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -796,7 +796,6 @@ static const struct flash_info spi_nor_i + { "s25sl032p", INFO(0x010215, 0x4d00, 64 * 1024, 64, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25sl064p", INFO(0x010216, 0x4d00, 64 * 1024, 128, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fs256s1", INFO6(0x010219, 0x4d0181, 64 * 1024, 512, 0)}, +- { "s25fs512s", INFO6(0x010220, 0x4d0081, 128 * 1024, 512, 0)}, + { "s25fl256s0", INFO(0x010219, 0x4d00, 256 * 1024, 128, 0) }, + { "s25fl256s1", INFO(0x010219, 0x4d01, 64 * 1024, 512, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fl512s", INFO(0x010220, 0x4d00, 256 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, +@@ -963,11 +962,9 @@ static int spansion_s25fs_disable_4kb_er + ret = nor->read_reg(nor, SPINOR_OP_SPANSION_RDAR, &cr3v, 1); + if (ret) + return ret; +-/* + if (!(cr3v & CR3V_4KB_ERASE_UNABLE)) + return -EPERM; + +-*/ + return 0; + } + +@@ -1334,6 +1331,8 @@ int spi_nor_scan(struct spi_nor *nor, co + + if (!mtd->name) + mtd->name = dev_name(dev); ++ if (info->name) ++ nor->vendor = info->name; + mtd->priv = nor; + mtd->type = MTD_NORFLASH; + mtd->writesize = 1; +--- a/include/linux/mtd/spi-nor.h ++++ b/include/linux/mtd/spi-nor.h +@@ -172,6 +172,7 @@ struct spi_nor { + bool sst_write_second; + u32 flags; + u8 cmd_buf[SPI_NOR_MAX_CMD_SIZE]; ++ char *vendor; + + int (*prepare)(struct spi_nor *nor, enum spi_nor_ops ops); + void (*unprepare)(struct spi_nor *nor, enum spi_nor_ops ops); diff --git a/target/linux/layerscape/patches-4.4/1111-mtd-spi-nor-disable-4kb-sector-erase-for-s25fl128.patch b/target/linux/layerscape/patches-4.4/1111-mtd-spi-nor-disable-4kb-sector-erase-for-s25fl128.patch new file mode 100644 index 0000000..503fa71 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1111-mtd-spi-nor-disable-4kb-sector-erase-for-s25fl128.patch @@ -0,0 +1,27 @@ +From 30d34abc80b0a602a1327bdfbddd42d250887049 Mon Sep 17 00:00:00 2001 +From: Yutang Jiang <yutang.jiang@nxp.com> +Date: Fri, 9 Sep 2016 22:56:12 +0800 +Subject: [PATCH 111/113] mtd: spi-nor: disable 4kb sector erase for s25fl128 + +As for s25fl128s flash, the sectors are organized either as a hybrid +combination of 4-kB and 64-kB sectors, or as uniform 256-kbyte sectors. +we should use the command 0xd8 to erase all bits, not the Parameter 4-kB +Sector Erase (P4E) command 0x20. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +Integrated-by: Yutang Jiang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/spi-nor.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +--- a/drivers/mtd/spi-nor/spi-nor.c ++++ b/drivers/mtd/spi-nor/spi-nor.c +@@ -802,7 +802,7 @@ static const struct flash_info spi_nor_i + { "s70fl01gs", INFO(0x010221, 0x4d00, 256 * 1024, 256, 0) }, + { "s25sl12800", INFO(0x012018, 0x0300, 256 * 1024, 64, 0) }, + { "s25sl12801", INFO(0x012018, 0x0301, 64 * 1024, 256, 0) }, +- { "s25fl128s", INFO6(0x012018, 0x4d0180, 64 * 1024, 256, SECT_4K | SPI_NOR_QUAD_READ ++ { "s25fl128s", INFO6(0x012018, 0x4d0180, 64 * 1024, 256, SPI_NOR_QUAD_READ + | SPI_NOR_DDR_QUAD_READ) }, + { "s25fl129p0", INFO(0x012018, 0x4d00, 256 * 1024, 64, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, + { "s25fl129p1", INFO(0x012018, 0x4d01, 64 * 1024, 256, SPI_NOR_DUAL_READ | SPI_NOR_QUAD_READ) }, diff --git a/target/linux/layerscape/patches-4.4/1112-driver-spi-fsl-quad-Hang-memcpy-Unhandled-fault-alig.patch b/target/linux/layerscape/patches-4.4/1112-driver-spi-fsl-quad-Hang-memcpy-Unhandled-fault-alig.patch new file mode 100644 index 0000000..c6311fc --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1112-driver-spi-fsl-quad-Hang-memcpy-Unhandled-fault-alig.patch @@ -0,0 +1,29 @@ +From f1b7824a42505669476f203e126fc26dd1006af2 Mon Sep 17 00:00:00 2001 +From: Yutang Jiang <yutang.jiang@nxp.com> +Date: Fri, 9 Sep 2016 22:57:55 +0800 +Subject: [PATCH 112/113] driver: spi: fsl-quad: Hang memcpy: Unhandled fault: + alignment fault + +vmap/iomap based on whether the buffer is in memory region or reserved region. +However, both map it as non-cacheable memory. +For armv8 specifically, non-cacheable mapping requests use a memory type +that has to be accessed aligned to the request size. memcpy() doesn't guarantee +that. memcpy_toio() can guarantee 4-bytes alignment. + +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +Integrated-by: Yutang Jiang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -1103,7 +1103,7 @@ static ssize_t fsl_qspi_read(struct spi_ + len); + + /* Read out the data directly from the AHB buffer.*/ +- memcpy(buf, q->ahb_addr + q->chip_base_addr + from - q->memmap_offs, ++ memcpy_toio(buf, q->ahb_addr + q->chip_base_addr + from - q->memmap_offs, + len); + + return len; diff --git a/target/linux/layerscape/patches-4.4/1113-mtd-spi-nor-fsl-quad-move-mtd_device_register-to-the.patch b/target/linux/layerscape/patches-4.4/1113-mtd-spi-nor-fsl-quad-move-mtd_device_register-to-the.patch new file mode 100644 index 0000000..f54f985 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/1113-mtd-spi-nor-fsl-quad-move-mtd_device_register-to-the.patch @@ -0,0 +1,49 @@ +From 6f1195d231ab576809fe2f4cff44a6e48cff2457 Mon Sep 17 00:00:00 2001 +From: Yutang Jiang <yutang.jiang@nxp.com> +Date: Fri, 2 Sep 2016 22:00:16 +0800 +Subject: [PATCH 113/113] mtd: spi-nor: fsl-quad: move mtd_device_register to + the last of probe + +After call mtd_device_register, the mtd devices should be workable immediately. +If before finish all of init work call the mtd_device_register, it will not +respond work request timely. + +For example, openwrt/lede have a AUTO split special flash partitions mechanism +while mtd driver register. So, before call mtd_device_register, must let all of +init work ready. + +Signed-off-by: Yutang Jiang <yutang.jiang@nxp.com> +--- + drivers/mtd/spi-nor/fsl-quadspi.c | 14 ++++++++++---- + 1 file changed, 10 insertions(+), 4 deletions(-) + +--- a/drivers/mtd/spi-nor/fsl-quadspi.c ++++ b/drivers/mtd/spi-nor/fsl-quadspi.c +@@ -1280,10 +1280,6 @@ static int fsl_qspi_probe(struct platfor + if (ret) + goto mutex_failed; + +- ret = mtd_device_register(mtd, NULL, 0); +- if (ret) +- goto mutex_failed; +- + /* Set the correct NOR size now. */ + if (q->nor_size == 0) { + q->nor_size = mtd->size; +@@ -1313,6 +1309,16 @@ static int fsl_qspi_probe(struct platfor + goto last_init_failed; + + fsl_qspi_clk_disable_unprep(q); ++ ++ for (i = 0; i < q->nor_num; i++) { ++ /* skip the holes */ ++ if (!q->has_second_chip) ++ i *= 2; ++ ++ ret = mtd_device_register(&q->nor[i].mtd, NULL, 0); ++ if (ret) ++ goto last_init_failed; ++ } + return 0; + + last_init_failed: diff --git a/target/linux/layerscape/patches-4.4/2119-armv8-aarch32-defconfig-Enable-CAAM-support.patch b/target/linux/layerscape/patches-4.4/2119-armv8-aarch32-defconfig-Enable-CAAM-support.patch new file mode 100644 index 0000000..567195d --- /dev/null +++ b/target/linux/layerscape/patches-4.4/2119-armv8-aarch32-defconfig-Enable-CAAM-support.patch @@ -0,0 +1,31 @@ +From 035fe1e511e053c6650f37626deb5da76dcc1d92 Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 11:29:48 +0800 +Subject: [PATCH 119/124] armv8: aarch32: defconfig: Enable CAAM support + +This patch is to enable the driver module for Freescale's Cryptographics +Accelerator and Assurance Module (CAAM) and related options. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/configs/ls_aarch32_defconfig | 9 +++++++-- + 1 file changed, 7 insertions(+), 2 deletions(-) + +--- a/arch/arm/configs/ls_aarch32_defconfig ++++ b/arch/arm/configs/ls_aarch32_defconfig +@@ -200,8 +200,13 @@ CONFIG_MAGIC_SYSRQ=y + # CONFIG_FTRACE is not set + CONFIG_PID_IN_CONTEXTIDR=y + CONFIG_CRYPTO_LZO=y +-# CONFIG_CRYPTO_ANSI_CPRNG is not set +-# CONFIG_CRYPTO_HW is not set ++CONFIG_CRYPTO_ANSI_CPRNG=y ++CONFIG_CRYPTO_DEV_FSL_CAAM=y ++CONFIG_ARM_CRYPTO=y ++CONFIG_CRYPTO_SHA1_ARM_NEON=y ++CONFIG_CRYPTO_SHA256_ARM=y ++CONFIG_CRYPTO_SHA512_ARM_NEON=y ++CONFIG_CRYPTO_AES_ARM_BS=y + CONFIG_CRC_CCITT=m + CONFIG_CRC_T10DIF=y + CONFIG_CRC7=m diff --git a/target/linux/layerscape/patches-4.4/2120-armv8-aarch32-defconfig-Enable-firmware-loading.patch b/target/linux/layerscape/patches-4.4/2120-armv8-aarch32-defconfig-Enable-firmware-loading.patch new file mode 100644 index 0000000..26a981c --- /dev/null +++ b/target/linux/layerscape/patches-4.4/2120-armv8-aarch32-defconfig-Enable-firmware-loading.patch @@ -0,0 +1,23 @@ +From 027ca2530ce94dd7d9954e57631aa34987db392e Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 11:38:36 +0800 +Subject: [PATCH 120/124] armv8: aarch32: defconfig: Enable firmware loading + +As some modules require userspace firmware loading support, such as +PPFE, add this feature in the defconfig for AArch32 on ARMv8. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/configs/ls_aarch32_defconfig | 1 - + 1 file changed, 1 deletion(-) + +--- a/arch/arm/configs/ls_aarch32_defconfig ++++ b/arch/arm/configs/ls_aarch32_defconfig +@@ -71,7 +71,6 @@ CONFIG_CAN=y + CONFIG_CAN_FLEXCAN=y + CONFIG_DEVTMPFS=y + CONFIG_DEVTMPFS_MOUNT=y +-# CONFIG_FW_LOADER is not set + CONFIG_DMA_CMA=y + CONFIG_MTD=y + CONFIG_MTD_CMDLINE_PARTS=y diff --git a/target/linux/layerscape/patches-4.4/2121-armv8-aarch32-defconfig-Enable-support-for-AHCI-SATA.patch b/target/linux/layerscape/patches-4.4/2121-armv8-aarch32-defconfig-Enable-support-for-AHCI-SATA.patch new file mode 100644 index 0000000..a3a67c3 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/2121-armv8-aarch32-defconfig-Enable-support-for-AHCI-SATA.patch @@ -0,0 +1,29 @@ +From ff37b165bdb100450c7996c9fac0fad2e6ffe31d Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 11:40:37 +0800 +Subject: [PATCH 121/124] armv8: aarch32: defconfig: Enable support for AHCI + SATA + +This patch is to enable support for the Freescale QorIQ AHCI SoC's +onboard AHCI SATA. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/configs/ls_aarch32_defconfig | 5 +++++ + 1 file changed, 5 insertions(+) + +--- a/arch/arm/configs/ls_aarch32_defconfig ++++ b/arch/arm/configs/ls_aarch32_defconfig +@@ -92,7 +92,12 @@ CONFIG_MTD_SPI_NOR=y + CONFIG_MTD_SPI_NOR_USE_4K_SECTORS=y + CONFIG_SPI_FSL_QUADSPI=y + CONFIG_BLK_DEV_SD=y ++CONFIG_CHR_DEV_SG=y + CONFIG_ATA=y ++CONFIG_SATA_AHCI=y ++CONFIG_SATA_AHCI_PLATFORM=y ++CONFIG_AHCI_QORIQ=y ++CONFIG_SATA_SIL24=y + CONFIG_BLK_DEV_LOOP=y + CONFIG_BLK_DEV_RAM=y + CONFIG_BLK_DEV_RAM_COUNT=8 diff --git a/target/linux/layerscape/patches-4.4/2122-armv8-aarch32-defconfig-Enable-USB-and-related-confi.patch b/target/linux/layerscape/patches-4.4/2122-armv8-aarch32-defconfig-Enable-USB-and-related-confi.patch new file mode 100644 index 0000000..83d372b --- /dev/null +++ b/target/linux/layerscape/patches-4.4/2122-armv8-aarch32-defconfig-Enable-USB-and-related-confi.patch @@ -0,0 +1,31 @@ +From 913db32774fe5c818112232823edfda1a706552f Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 11:42:40 +0800 +Subject: [PATCH 122/124] armv8: aarch32: defconfig: Enable USB and related + configuration options + +This patch is to enable USB and related configuration options for +AArch32 on ARMv8. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/configs/ls_aarch32_defconfig | 6 ++++++ + 1 file changed, 6 insertions(+) + +--- a/arch/arm/configs/ls_aarch32_defconfig ++++ b/arch/arm/configs/ls_aarch32_defconfig +@@ -154,8 +154,14 @@ CONFIG_MFD_SYSCON=y + CONFIG_FB=y + CONFIG_USB=y + CONFIG_USB_XHCI_HCD=y ++CONFIG_USB_EHCI_HCD=y ++CONFIG_USB_EHCI_HCD_PLATFORM=y ++CONFIG_USB_OHCI_HCD=y ++CONFIG_USB_OHCI_HCD_PLATFORM=y ++CONFIG_USB_STORAGE=y + CONFIG_USB_DWC3=y + CONFIG_USB_STORAGE=y ++CONFIG_USB_ULPI=y + CONFIG_MMC=y + CONFIG_MMC_SDHCI=y + CONFIG_MMC_SDHCI_PLTFM=y diff --git a/target/linux/layerscape/patches-4.4/2123-armv8-aarch32-defconfig-Enable-KVM-related-configura.patch b/target/linux/layerscape/patches-4.4/2123-armv8-aarch32-defconfig-Enable-KVM-related-configura.patch new file mode 100644 index 0000000..4556ad9 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/2123-armv8-aarch32-defconfig-Enable-KVM-related-configura.patch @@ -0,0 +1,59 @@ +From dbf356cd062d6313f90323a77d4cf7c820dae40f Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 13:51:32 +0800 +Subject: [PATCH 123/124] armv8: aarch32: defconfig: Enable KVM-related + configuration options + +This patch is to enable KVM-related configuration options for host and +guest. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/configs/ls_aarch32_defconfig | 10 ++++++++++ + 1 file changed, 10 insertions(+) + +--- a/arch/arm/configs/ls_aarch32_defconfig ++++ b/arch/arm/configs/ls_aarch32_defconfig +@@ -102,7 +102,11 @@ CONFIG_BLK_DEV_LOOP=y + CONFIG_BLK_DEV_RAM=y + CONFIG_BLK_DEV_RAM_COUNT=8 + CONFIG_BLK_DEV_RAM_SIZE=262144 ++CONFIG_VIRTIO_BLK=y + CONFIG_NETDEVICES=y ++CONFIG_TUN=y ++CONFIG_VIRTIO_NET=y ++CONFIG_VHOST_NET=y + CONFIG_NET_VENDOR_FREESCALE is not set + CONFIG_FSL_BMAN=y + CONFIG_FSL_QMAN=y +@@ -130,6 +134,8 @@ CONFIG_SERIAL_8250=y + CONFIG_SERIAL_8250_CONSOLE=y + CONFIG_SERIAL_8250_EXTENDED=y + CONFIG_SERIAL_8250_SHARE_IRQ=y ++CONFIG_SERIAL_AMBA_PL011=y ++CONFIG_SERIAL_AMBA_PL011_CONSOLE=y + CONFIG_SERIAL_OF_PLATFORM=y + CONFIG_SERIAL_FSL_LPUART=y + CONFIG_SERIAL_FSL_LPUART_CONSOLE=y +@@ -170,6 +176,7 @@ CONFIG_RTC_CLASS=y + CONFIG_RTC_DRV_DS3232=y + CONFIG_DMADEVICES=y + CONFIG_FSL_EDMA=y ++CONFIG_VIRTIO_PCI=y + CONFIG_STAGING=y + CONFIG_CLK_QORIQ=y + # CONFIG_IOMMU_SUPPORT is not set +@@ -191,6 +198,7 @@ CONFIG_VFAT_FS=y + CONFIG_NTFS_FS=m + CONFIG_TMPFS=y + CONFIG_TMPFS_POSIX_ACL=y ++CONFIG_HUGETLBFS=y + CONFIG_CONFIGFS_FS=y + CONFIG_JFFS2_FS=y + CONFIG_NFS_FS=y +@@ -221,3 +229,5 @@ CONFIG_CRC_CCITT=m + CONFIG_CRC_T10DIF=y + CONFIG_CRC7=m + CONFIG_LIBCRC32C=m ++CONFIG_VIRTUALIZATION=y ++CONFIG_KVM=y diff --git a/target/linux/layerscape/patches-4.4/2124-armv8-aarch32-defconfig-Enable-FTM-alarm-support.patch b/target/linux/layerscape/patches-4.4/2124-armv8-aarch32-defconfig-Enable-FTM-alarm-support.patch new file mode 100644 index 0000000..ae05a88 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/2124-armv8-aarch32-defconfig-Enable-FTM-alarm-support.patch @@ -0,0 +1,23 @@ +From 6df7fdf7e0c76df7acc2d1a3a287bf094a94c4ff Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 13:55:44 +0800 +Subject: [PATCH 124/124] armv8: aarch32: defconfig: Enable FTM alarm support + +This patch is to enable FTM alarm support. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/configs/ls_aarch32_defconfig | 2 ++ + 1 file changed, 2 insertions(+) + +--- a/arch/arm/configs/ls_aarch32_defconfig ++++ b/arch/arm/configs/ls_aarch32_defconfig +@@ -181,6 +181,8 @@ CONFIG_STAGING=y + CONFIG_CLK_QORIQ=y + # CONFIG_IOMMU_SUPPORT is not set + CONFIG_MEMORY=y ++CONFIG_LS_SOC_DRIVERS=y ++CONFIG_FTM_ALARM=y + CONFIG_PWM=y + CONFIG_PWM_FSL_FTM=y + # CONFIG_RESET_CONTROLLER is not set diff --git a/target/linux/layerscape/patches-4.4/3071-arm64-dts-add-device-tree-for-ls1012a-SoC-and-boards.patch b/target/linux/layerscape/patches-4.4/3071-arm64-dts-add-device-tree-for-ls1012a-SoC-and-boards.patch new file mode 100644 index 0000000..10c11eb --- /dev/null +++ b/target/linux/layerscape/patches-4.4/3071-arm64-dts-add-device-tree-for-ls1012a-SoC-and-boards.patch @@ -0,0 +1,880 @@ +From 70e0080366e76dabf90b713f57fb9fc47aa35557 Mon Sep 17 00:00:00 2001 +From: Yangbo Lu <yangbo.lu@nxp.com> +Date: Thu, 11 Aug 2016 10:36:05 +0800 +Subject: [PATCH 071/113] arm64: dts: add device tree for ls1012a SoC and + boards +MIME-Version: 1.0 +Content-Type: text/plain; charset=UTF-8 +Content-Transfer-Encoding: 8bit + +This patch is to add device tree for ls1012a SoC and RDB/FREEDOM boards. + +Signed-off-by: Pratiyush Mohan Srivastava <pratiyush.srivastava@nxp.com> +Signed-off-by: Prabhakar Kushwaha <prabhakar.kushwaha@nxp.com> +Signed-off-by: Yunhui Cui <yunhui.cui@nxp.com> +Signed-off-by: Rajesh Bhagat <rajesh.bhagat@nxp.com> +Signed-off-by: Alison Wang <alison.wang@nxp.com> +Signed-off-by: Horia Geantă <horia.geanta@nxp.com> +Signed-off-by: Bhaskar Upadhaya <Bhaskar.Upadhaya@freescale.com> +Signed-off-by: Tang Yuantian <yuantian.tang@nxp.com> +Signed-off-by: Chenhui Zhao <chenhui.zhao@nxp.com> +Signed-off-by: Jia Hongtao <hongtao.jia@nxp.com> +Signed-off-by: Calvin Johnson <calvin.johnson@nxp.com> +[yangbolu: integrate] +Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com> +--- + arch/arm64/boot/dts/freescale/Makefile | 2 + + arch/arm64/boot/dts/freescale/fsl-ls1012a-frdm.dts | 186 +++++++ + arch/arm64/boot/dts/freescale/fsl-ls1012a-rdb.dts | 114 +++++ + arch/arm64/boot/dts/freescale/fsl-ls1012a.dtsi | 526 ++++++++++++++++++++ + 4 files changed, 828 insertions(+) + create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-frdm.dts + create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a-rdb.dts + create mode 100644 arch/arm64/boot/dts/freescale/fsl-ls1012a.dtsi + +--- a/arch/arm64/boot/dts/freescale/Makefile ++++ b/arch/arm64/boot/dts/freescale/Makefile +@@ -2,6 +2,8 @@ dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2 + dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2080a-rdb.dtb + dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls2080a-simu.dtb + dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1043a-rdb.dtb ++dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-rdb.dtb ++dtb-$(CONFIG_ARCH_LAYERSCAPE) += fsl-ls1012a-frdm.dtb + + always := $(dtb-y) + subdir-y := $(dts-dirs) +--- /dev/null ++++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-frdm.dts +@@ -0,0 +1,186 @@ ++/* ++ * Device Tree Include file for Freescale Layerscape-1012A family SoC. ++ * ++ * Copyright 2016, Freescale Semiconductor Inc. ++ ++ * Redistribution and use in source and binary forms, with or without ++ * modification, are permitted provided that the following conditions are met: ++ * * Redistributions of source code must retain the above copyright ++ * notice, this list of conditions and the following disclaimer. ++ * * Redistributions in binary form must reproduce the above copyright ++ * notice, this list of conditions and the following disclaimer in the ++ * documentation and/or other materials provided with the distribution. ++ * * Neither the name of Freescale Semiconductor nor the ++ * names of its contributors may be used to endorse or promote products ++ * derived from this software without specific prior written permission. ++ * ++ * ++ * ALTERNATIVELY, this software may be distributed under the terms of the ++ * GNU General Public License ("GPL") as published by the Free Software ++ * Foundation, either version 2 of that License or (at your option) any ++ * later version. ++ * ++ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY ++ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED ++ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE ++ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY ++ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES ++ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; ++ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ++ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ++ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS ++ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ++ */ ++/dts-v1/; ++ ++#include "fsl-ls1012a.dtsi" ++ ++/ { ++ model = "LS1012A FREEDOM Board"; ++ compatible = "fsl,ls1012a-frdm", "fsl,ls1012a"; ++ ++ aliases { ++ crypto = &crypto; ++ }; ++ ++ sys_mclk: clock-mclk { ++ compatible = "fixed-clock"; ++ #clock-cells = <0>; ++ clock-frequency = <25000000>; ++ }; ++ ++ regulators { ++ compatible = "simple-bus"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ ++ reg_1p8v: regulator@0 { ++ compatible = "regulator-fixed"; ++ reg = <0>; ++ regulator-name = "1P8V"; ++ regulator-min-microvolt = <1800000>; ++ regulator-max-microvolt = <1800000>; ++ regulator-always-on; ++ }; ++ }; ++ ++ sound { ++ compatible = "simple-audio-card"; ++ simple-audio-card,format = "i2s"; ++ simple-audio-card,widgets = ++ "Microphone", "Microphone Jack", ++ "Headphone", "Headphone Jack", ++ "Speaker", "Speaker Ext", ++ "Line", "Line In Jack"; ++ simple-audio-card,routing = ++ "MIC_IN", "Microphone Jack", ++ "Microphone Jack", "Mic Bias", ++ "LINE_IN", "Line In Jack", ++ "Headphone Jack", "HP_OUT", ++ "Speaker Ext", "LINE_OUT"; ++ ++ simple-audio-card,cpu { ++ sound-dai = <&sai2>; ++ frame-master; ++ bitclock-master; ++ }; ++ ++ simple-audio-card,codec { ++ sound-dai = <&codec>; ++ frame-master; ++ bitclock-master; ++ system-clock-frequency = <25000000>; ++ }; ++ }; ++}; ++ ++&qspi { ++ num-cs = <2>; ++ bus-num = <0>; ++ status = "okay"; ++ fsl,ddr-sampling-point = <4>; ++ ++ qflash0: s25fs512s@0 { ++ compatible = "spansion,m25p80"; ++ #address-cells = <1>; ++ #size-cells = <1>; ++ spi-max-frequency = <20000000>; ++ m25p,fast-read; ++ reg = <0>; ++ }; ++}; ++&ftm0 { ++ status = "okay"; ++}; ++ ++&i2c0 { ++ status = "okay"; ++ ++ codec: sgtl5000@a { ++ #sound-dai-cells = <0>; ++ compatible = "fsl,sgtl5000"; ++ reg = <0xa>; ++ VDDA-supply = <®_1p8v>; ++ VDDIO-supply = <®_1p8v>; ++ clocks = <&sys_mclk 1>; ++ }; ++}; ++ ++&duart0 { ++ status = "okay"; ++}; ++&pfe { ++ status = "okay"; ++ ethernet@0 { ++ compatible = "fsl,pfe-gemac-port"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = < 0x0 >; /* GEM_ID */ ++ fsl,gemac-bus-id = <0x0>; /* BUS_ID */ ++ fsl,gemac-phy-id = <0x2>; /* PHY_ID */ ++ fsl,mdio-mux-val = <0x0>; ++ local-mac-address = [ 00 1A 2B 3C 4D 5E ]; ++ phy-mode = "sgmii"; ++ fsl,pfe-gemac-if-name = "eth0"; ++ fsl,pfe-phy-if-flags = <0x0>; ++ fsl,pfe-gemac-mode = <0x1B00>; /* GEMAC_SW_CONF | GEMAC_SW_FULL_DUPLEX | GEMAC_SW_SPEED_1G */ ++ ++ mdio@0 { ++ reg = <0x1>; /* enabled/disabled */ ++ fsl,mdio-phy-mask = <0xFFFFFFF9>; ++ }; ++ }; ++ ethernet@1 { ++ compatible = "fsl,pfe-gemac-port"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = < 0x1 >; /* GEM_ID */ ++ fsl,gemac-bus-id = < 0x1 >; /* BUS_ID */ ++ fsl,gemac-phy-id = < 0x1 >; /* PHY_ID */ ++ fsl,mdio-mux-val = <0x0>; ++ local-mac-address = [ 00 AA BB CC DD EE ]; ++ phy-mode = "sgmii"; ++ fsl,pfe-gemac-if-name = "eth1"; ++ fsl,pfe-phy-if-flags = <0x0>; ++ fsl,pfe-gemac-mode = <0x1B00>; /* GEMAC_SW_CONF | GEMAC_SW_FULL_DUPLEX | GEMAC_SW_SPEED_1G */ ++ mdio@0 { ++ reg = <0x0>; /* enabled/disabled */ ++ fsl,mdio-phy-mask = <0xFFFFFFF9>; ++ }; ++ ++ }; ++ ++}; ++ ++ ++&esdhc0 { ++ status = "disabled"; ++}; ++ ++&esdhc1 { ++ status = "disabled"; ++}; ++ ++&sai2 { ++ status = "okay"; ++}; +--- /dev/null ++++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a-rdb.dts +@@ -0,0 +1,114 @@ ++/* ++ * Device Tree Include file for Freescale Layerscape-1012A family SoC. ++ * ++ * Copyright 2016, Freescale Semiconductor Inc. ++ ++ * Redistribution and use in source and binary forms, with or without ++ * modification, are permitted provided that the following conditions are met: ++ * * Redistributions of source code must retain the above copyright ++ * notice, this list of conditions and the following disclaimer. ++ * * Redistributions in binary form must reproduce the above copyright ++ * notice, this list of conditions and the following disclaimer in the ++ * documentation and/or other materials provided with the distribution. ++ * * Neither the name of Freescale Semiconductor nor the ++ * names of its contributors may be used to endorse or promote products ++ * derived from this software without specific prior written permission. ++ * ++ * ++ * ALTERNATIVELY, this software may be distributed under the terms of the ++ * GNU General Public License ("GPL") as published by the Free Software ++ * Foundation, either version 2 of that License or (at your option) any ++ * later version. ++ * ++ * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY ++ * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED ++ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE ++ * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY ++ * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES ++ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; ++ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ++ * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ++ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS ++ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ++ */ ++/dts-v1/; ++ ++#include "fsl-ls1012a.dtsi" ++ ++/ { ++ model = "LS1012A RDB Board"; ++ compatible = "fsl,ls1012a-rdb", "fsl,ls1012a"; ++ ++ aliases { ++ crypto = &crypto; ++ }; ++}; ++ ++&qspi { ++ num-cs = <2>; ++ bus-num = <0>; ++ status = "okay"; ++ fsl,ddr-sampling-point = <4>; ++ ++ qflash0: s25fs512s@0 { ++ compatible = "spansion,m25p80"; ++ #address-cells = <1>; ++ #size-cells = <1>; ++ spi-max-frequency = <20000000>; ++ m25p,fast-read; ++ reg = <0>; ++ }; ++}; ++&ftm0 { ++ status = "okay"; ++}; ++ ++&i2c0 { ++ status = "okay"; ++}; ++ ++&duart0 { ++ status = "okay"; ++}; ++&pfe { ++ status = "okay"; ++ ethernet@0 { ++ compatible = "fsl,pfe-gemac-port"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = < 0x0 >; /* GEM_ID */ ++ fsl,gemac-bus-id = <0x0>; /* BUS_ID */ ++ fsl,gemac-phy-id = <0x2>; /* PHY_ID */ ++ fsl,mdio-mux-val = <0x0>; ++ local-mac-address = [ 00 1A 2B 3C 4D 5E ]; ++ phy-mode = "sgmii"; ++ fsl,pfe-gemac-if-name = "eth0"; ++ fsl,pfe-phy-if-flags = <0x0>; ++ fsl,pfe-gemac-mode = <0x1B00>; /* GEMAC_SW_CONF | GEMAC_SW_FULL_DUPLEX | GEMAC_SW_SPEED_1G */ ++ ++ mdio@0 { ++ reg = <0x1>; /* enabled/disabled */ ++ fsl,mdio-phy-mask = <0xFFFFFFF9>; ++ }; ++ }; ++ ethernet@1 { ++ compatible = "fsl,pfe-gemac-port"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = < 0x1 >; /* GEM_ID */ ++ fsl,gemac-bus-id = < 0x1 >; /* BUS_ID */ ++ fsl,gemac-phy-id = < 0x1 >; /* PHY_ID */ ++ fsl,mdio-mux-val = <0x0>; ++ local-mac-address = [ 00 AA BB CC DD EE ]; ++ phy-mode = "rgmii"; ++ fsl,pfe-gemac-if-name = "eth2"; ++ fsl,pfe-phy-if-flags = <0x0>; ++ fsl,pfe-gemac-mode = <0x1B00>; /* GEMAC_SW_CONF | GEMAC_SW_FULL_DUPLEX | GEMAC_SW_SPEED_1G */ ++ mdio@0 { ++ reg = <0x0>; /* enabled/disabled */ ++ fsl,mdio-phy-mask = <0xFFFFFFF9>; ++ }; ++ ++ }; ++ ++}; +--- /dev/null ++++ b/arch/arm64/boot/dts/freescale/fsl-ls1012a.dtsi +@@ -0,0 +1,526 @@ ++/* ++ * Device Tree Include file for Freescale Layerscape-1043A family SoC. ++ * ++ * Copyright 2016, Freescale Semiconductor ++ * ++ * This file is dual-licensed: you can use it either under the terms ++ * of the GPLv2 or the X11 license, at your option. Note that this dual ++ * licensing only applies to this file, and not this project as a ++ * whole. ++ * ++ * a) This library is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License as ++ * published by the Free Software Foundation; either version 2 of the ++ * License, or (at your option) any later version. ++ * ++ * This library is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * Or, alternatively, ++ * ++ * b) Permission is hereby granted, free of charge, to any person ++ * obtaining a copy of this software and associated documentation ++ * files (the "Software"), to deal in the Software without ++ * restriction, including without limitation the rights to use, ++ * copy, modify, merge, publish, distribute, sublicense, and/or ++ * sell copies of the Software, and to permit persons to whom the ++ * Software is furnished to do so, subject to the following ++ * conditions: ++ * ++ * The above copyright notice and this permission notice shall be ++ * included in all copies or substantial portions of the Software. ++ * ++ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, ++ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES ++ * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND ++ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT ++ * HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, ++ * WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING ++ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR ++ * OTHER DEALINGS IN THE SOFTWARE. ++ */ ++ ++#include <dt-bindings/thermal/thermal.h> ++ ++/ { ++ compatible = "fsl,ls1012a"; ++ interrupt-parent = <&gic>; ++ #address-cells = <2>; ++ #size-cells = <2>; ++ ++ cpus { ++ #address-cells = <1>; ++ #size-cells = <0>; ++ ++ /* ++ * We expect the enable-method for cpu's to be "psci", but this ++ * is dependent on the SoC FW, which will fill this in. ++ * ++ * Currently supported enable-method is psci v0.2 ++ */ ++ cpu0: cpu@0 { ++ device_type = "cpu"; ++ compatible = "arm,cortex-a53"; ++ reg = <0x0>; ++ clocks = <&clockgen 1 0>; ++ #cooling-cells = <2>; ++ }; ++ ++ }; ++ ++ ++ sysclk: sysclk { ++ compatible = "fixed-clock"; ++ #clock-cells = <0>; ++ clock-frequency = <100000000>; ++ clock-output-names = "sysclk"; ++ }; ++ ++ timer { ++ compatible = "arm,armv8-timer"; ++ interrupts = <1 13 0x1>, /* Physical Secure PPI */ ++ <1 14 0x1>, /* Physical Non-Secure PPI */ ++ <1 11 0x1>, /* Virtual PPI */ ++ <1 10 0x1>; /* Hypervisor PPI */ ++ arm,reread-timer; ++ }; ++ ++ pmu { ++ compatible = "arm,armv8-pmuv3"; ++ interrupts = <0 106 0x4>; ++ }; ++ ++ gic: interrupt-controller@1400000 { ++ compatible = "arm,gic-400"; ++ #interrupt-cells = <3>; ++ interrupt-controller; ++ reg = <0x0 0x1401000 0 0x1000>, /* GICD */ ++ <0x0 0x1402000 0 0x2000>, /* GICC */ ++ <0x0 0x1404000 0 0x2000>, /* GICH */ ++ <0x0 0x1406000 0 0x2000>; /* GICV */ ++ interrupts = <1 9 0xf08>; ++ }; ++ ++ soc { ++ compatible = "simple-bus"; ++ #address-cells = <2>; ++ #size-cells = <2>; ++ ranges; ++ ++ clockgen: clocking@1ee1000 { ++ compatible = "fsl,ls1012a-clockgen"; ++ reg = <0x0 0x1ee1000 0x0 0x1000>; ++ #clock-cells = <2>; ++ clocks = <&sysclk>; ++ }; ++ ++ scfg: scfg@1570000 { ++ compatible = "fsl,ls1012a-scfg", ++ "fsl,ls1043a-scfg", ++ "syscon"; ++ reg = <0x0 0x1570000 0x0 0x10000>; ++ big-endian; ++ }; ++ ++ crypto: crypto@1700000 { ++ compatible = "fsl,sec-v5.4", "fsl,sec-v5.0", ++ "fsl,sec-v4.0"; ++ fsl,sec-era = <8>; ++ #address-cells = <1>; ++ #size-cells = <1>; ++ ranges = <0x0 0x00 0x1700000 0x100000>; ++ reg = <0x00 0x1700000 0x0 0x100000>; ++ interrupts = <0 75 0x4>; ++ ++ sec_jr0: jr@10000 { ++ compatible = "fsl,sec-v5.4-job-ring", ++ "fsl,sec-v5.0-job-ring", ++ "fsl,sec-v4.0-job-ring"; ++ reg = <0x10000 0x10000>; ++ interrupts = <0 71 0x4>; ++ }; ++ ++ sec_jr1: jr@20000 { ++ compatible = "fsl,sec-v5.4-job-ring", ++ "fsl,sec-v5.0-job-ring", ++ "fsl,sec-v4.0-job-ring"; ++ reg = <0x20000 0x10000>; ++ interrupts = <0 72 0x4>; ++ }; ++ ++ sec_jr2: jr@30000 { ++ compatible = "fsl,sec-v5.4-job-ring", ++ "fsl,sec-v5.0-job-ring", ++ "fsl,sec-v4.0-job-ring"; ++ reg = <0x30000 0x10000>; ++ interrupts = <0 73 0x4>; ++ }; ++ ++ sec_jr3: jr@40000 { ++ compatible = "fsl,sec-v5.4-job-ring", ++ "fsl,sec-v5.0-job-ring", ++ "fsl,sec-v4.0-job-ring"; ++ reg = <0x40000 0x10000>; ++ interrupts = <0 74 0x4>; ++ }; ++ }; ++ ++ ++ dcfg: dcfg@1ee0000 { ++ compatible = "fsl,ls1012a-dcfg", ++ "fsl,ls1043a-dcfg", ++ "syscon"; ++ reg = <0x0 0x1ee0000 0x0 0x10000>; ++ }; ++ ++ reset: reset@1EE00B0 { ++ compatible = "fsl,ls-reset"; ++ reg = <0x0 0x1EE00B0 0x0 0x4>; ++ big-endian; ++ }; ++ ++ rcpm: rcpm@1ee2000 { ++ compatible = "fsl,ls1012a-rcpm", ++ "fsl,ls1043a-rcpm", ++ "fsl,qoriq-rcpm-2.1"; ++ reg = <0x0 0x1ee2000 0x0 0x10000>; ++ }; ++ ++ ftm0: ftm0@29d0000 { ++ compatible = "fsl,ftm-alarm"; ++ reg = <0x0 0x29d0000 0x0 0x10000>; ++ interrupts = <0 86 0x4>; ++ big-endian; ++ rcpm-wakeup = <&rcpm 0x00020000 0x0>; ++ status = "okay"; ++ }; ++ ++ esdhc0: esdhc@1560000 { ++ compatible = "fsl,ls1012a-esdhc0", "fsl,esdhc"; ++ reg = <0x0 0x1560000 0x0 0x10000>; ++ interrupts = <0 62 0x4>; ++ clock-frequency = <0>; ++ voltage-ranges = <1800 1800 3300 3300>; ++ sdhci,auto-cmd12; ++ big-endian; ++ bus-width = <4>; ++ }; ++ ++ esdhc1: esdhc@1580000 { ++ compatible = "fsl,ls1012a-esdhc1", "fsl,esdhc"; ++ reg = <0x0 0x1580000 0x0 0x10000>; ++ interrupts = <0 65 0x4>; ++ clock-frequency = <0>; ++ voltage-ranges = <1800 1800 3300 3300>; ++ sdhci,auto-cmd12; ++ big-endian; ++ bus-width = <4>; ++ }; ++ ++ dspi0: dspi@2100000 { ++ compatible = "fsl,ls1012a-dspi", ++ "fsl,ls1043a-dspi", ++ "fsl,ls1021a-v1.0-dspi"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = <0x0 0x2100000 0x0 0x10000>; ++ interrupts = <0 64 0x4>; ++ clock-names = "dspi"; ++ clocks = <&clockgen 4 0>; ++ spi-num-chipselects = <5>; ++ big-endian; ++ status = "enabled"; ++ }; ++ ++ qspi: quadspi@1550000 { ++ compatible = "fsl,ls1012a-qspi", ++ "fsl,ls1043a-qspi", ++ "fsl,ls1021a-qspi"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = <0x0 0x1550000 0x0 0x10000>, ++ <0x0 0x40000000 0x0 0x4000000>; ++ reg-names = "QuadSPI", "QuadSPI-memory"; ++ interrupts = <0 99 0x4>; ++ clock-names = "qspi_en", "qspi"; ++ clocks = <&clockgen 4 0>, <&clockgen 4 0>; ++ big-endian; ++ amba-base = <0x42000000>; ++ }; ++ ++ tmu: tmu@1f00000 { ++ compatible = "fsl,qoriq-tmu", "fsl,ls1012a-tmu"; ++ reg = <0x0 0x1f00000 0x0 0x10000>; ++ interrupts = <0 33 0x4>; ++ fsl,tmu-range = <0xb0000 0x9002a 0x6004c 0x30062>; ++ fsl,tmu-calibration = <0x00000000 0x00000026 ++ 0x00000001 0x0000002d ++ 0x00000002 0x00000032 ++ 0x00000003 0x00000039 ++ 0x00000004 0x0000003f ++ 0x00000005 0x00000046 ++ 0x00000006 0x0000004d ++ 0x00000007 0x00000054 ++ 0x00000008 0x0000005a ++ 0x00000009 0x00000061 ++ 0x0000000a 0x0000006a ++ 0x0000000b 0x00000071 ++ ++ 0x00010000 0x00000025 ++ 0x00010001 0x0000002c ++ 0x00010002 0x00000035 ++ 0x00010003 0x0000003d ++ 0x00010004 0x00000045 ++ 0x00010005 0x0000004e ++ 0x00010006 0x00000057 ++ 0x00010007 0x00000061 ++ 0x00010008 0x0000006b ++ 0x00010009 0x00000076 ++ ++ 0x00020000 0x00000029 ++ 0x00020001 0x00000033 ++ 0x00020002 0x0000003d ++ 0x00020003 0x00000049 ++ 0x00020004 0x00000056 ++ 0x00020005 0x00000061 ++ 0x00020006 0x0000006d ++ ++ 0x00030000 0x00000021 ++ 0x00030001 0x0000002a ++ 0x00030002 0x0000003c ++ 0x00030003 0x0000004e>; ++ big-endian; ++ #thermal-sensor-cells = <1>; ++ }; ++ ++ thermal-zones { ++ cpu_thermal: cpu-thermal { ++ polling-delay-passive = <1000>; ++ polling-delay = <5000>; ++ ++ thermal-sensors = <&tmu 0>; ++ ++ trips { ++ cpu_alert: cpu-alert { ++ temperature = <85000>; ++ hysteresis = <2000>; ++ type = "passive"; ++ }; ++ cpu_crit: cpu-crit { ++ temperature = <95000>; ++ hysteresis = <2000>; ++ type = "critical"; ++ }; ++ }; ++ ++ cooling-maps { ++ map0 { ++ trip = <&cpu_alert>; ++ cooling-device = ++ <&cpu0 THERMAL_NO_LIMIT ++ THERMAL_NO_LIMIT>; ++ }; ++ }; ++ }; ++ }; ++ ++ i2c0: i2c@2180000 { ++ compatible = "fsl,vf610-i2c"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = <0x0 0x2180000 0x0 0x10000>; ++ interrupts = <0 56 0x4>; ++ clock-names = "i2c"; ++ clocks = <&clockgen 4 0>; ++ status = "disabled"; ++ }; ++ ++ i2c1: i2c@2190000 { ++ compatible = "fsl,vf610-i2c"; ++ #address-cells = <1>; ++ #size-cells = <0>; ++ reg = <0x0 0x2190000 0x0 0x10000>; ++ interrupts = <0 57 0x4>; ++ clock-names = "i2c"; ++ clocks = <&clockgen 4 0>; ++ status = "disabled"; ++ }; ++ ++ ++ duart0: serial@21c0500 { ++ compatible = "fsl,ns16550", "ns16550a"; ++ reg = <0x00 0x21c0500 0x0 0x100>; ++ interrupts = <0 54 0x4>; ++ clocks = <&clockgen 4 0>; ++ }; ++ ++ duart1: serial@21c0600 { ++ compatible = "fsl,ns16550", "ns16550a"; ++ reg = <0x00 0x21c0600 0x0 0x100>; ++ interrupts = <0 54 0x4>; ++ clocks = <&clockgen 4 0>; ++ }; ++ ++ gpio0: gpio@2300000 { ++ compatible = "fsl,qoriq-gpio"; ++ reg = <0x0 0x2300000 0x0 0x10000>; ++ interrupts = <0 66 0x4>; ++ gpio-controller; ++ #gpio-cells = <2>; ++ interrupt-controller; ++ #interrupt-cells = <2>; ++ }; ++ ++ gpio1: gpio@2310000 { ++ compatible = "fsl,qoriq-gpio"; ++ reg = <0x0 0x2310000 0x0 0x10000>; ++ interrupts = <0 67 0x4>; ++ gpio-controller; ++ #gpio-cells = <2>; ++ interrupt-controller; ++ #interrupt-cells = <2>; ++ }; ++ ++ wdog0: wdog@2ad0000 { ++ compatible = "fsl,ls1012a-wdt", ++ "fsl,ls1043a-wdt", ++ "fsl,imx21-wdt"; ++ reg = <0x0 0x2ad0000 0x0 0x10000>; ++ interrupts = <0 83 0x4>; ++ clocks = <&clockgen 4 0>; ++ clock-names = "wdog"; ++ big-endian; ++ }; ++ ++ sai1: sai@2b50000 { ++ #sound-dai-cells = <0>; ++ compatible = "fsl,vf610-sai"; ++ reg = <0x0 0x2b50000 0x0 0x10000>; ++ interrupts = <0 148 0x4>; ++ clocks = <&clockgen 4 3>, <&clockgen 4 3>, ++ <&clockgen 4 3>, <&clockgen 4 3>; ++ clock-names = "bus", "mclk1", "mclk2", "mclk3"; ++ dma-names = "tx", "rx"; ++ dmas = <&edma0 1 47>, ++ <&edma0 1 46>; ++ status = "disabled"; ++ }; ++ ++ sai2: sai@2b60000 { ++ #sound-dai-cells = <0>; ++ compatible = "fsl,vf610-sai"; ++ reg = <0x0 0x2b60000 0x0 0x10000>; ++ interrupts = <0 149 0x4>; ++ clocks = <&clockgen 4 3>, <&clockgen 4 3>, ++ <&clockgen 4 3>, <&clockgen 4 3>; ++ clock-names = "bus", "mclk1", "mclk2", "mclk3"; ++ dma-names = "tx", "rx"; ++ dmas = <&edma0 1 45>, ++ <&edma0 1 44>; ++ status = "disabled"; ++ }; ++ ++ edma0: edma@2c00000 { ++ #dma-cells = <2>; ++ compatible = "fsl,vf610-edma"; ++ reg = <0x0 0x2c00000 0x0 0x10000>, ++ <0x0 0x2c10000 0x0 0x10000>, ++ <0x0 0x2c20000 0x0 0x10000>; ++ interrupts = <0 103 0x4>, ++ <0 103 0x4>; ++ interrupt-names = "edma-tx", "edma-err"; ++ dma-channels = <32>; ++ big-endian; ++ clock-names = "dmamux0", "dmamux1"; ++ clocks = <&clockgen 4 3>, ++ <&clockgen 4 3>; ++ }; ++ ++ sata: sata@3200000 { ++ compatible = "fsl,ls1012a-ahci"; ++ reg = <0x0 0x3200000 0x0 0x10000>; ++ interrupts = <0 69 0x4>; ++ clocks = <&clockgen 4 0>; ++ }; ++ ++ msi2: msi-controller2@1572000 { ++ compatible ="fsl,1s1012a-msi", "fsl,1s1021a-msi"; ++ reg = <0x0 0x1572000 0x0 0x4>, ++ <0x0 0x1572004 0x0 0x4>; ++ reg-names = "msiir", "msir"; ++ msi-controller; ++ interrupts = <0 126 0x4>; ++ }; ++ ++ usb@8600000 { ++ compatible = "fsl-usb2-dr-v2.5", "fsl-usb2-dr"; ++ reg = <0x0 0x8600000 0x0 0x1000>; ++ interrupts = <0 139 0x4>; ++ dr_mode = "host"; ++ phy_type = "ulpi"; ++ fsl,usb-erratum-a005697; ++ }; ++ ++ usb0: usb3@2f00000 { ++ compatible = "snps,dwc3"; ++ reg = <0x0 0x2f00000 0x0 0x10000>; ++ interrupts = <0 60 0x4>; ++ dr_mode = "host"; ++ configure-gfladj; ++ snps,dis_rxdet_inp3_quirk; ++ }; ++ ++ pcie@3400000 { ++ compatible = "fsl,ls1012a-pcie", ++ "fsl,ls1043a-pcie", ++ "snps,dw-pcie"; ++ reg = <0x00 0x03400000 0x0 0x00100000 /* controller registers */ ++ 0x40 0x00000000 0x0 0x00002000>; /* configuration space */ ++ reg-names = "regs", "config"; ++ interrupts = <0 118 0x4>, /* controller interrupt */ ++ <0 117 0x4>; /* PME interrupt */ ++ interrupt-names = "intr", "pme"; ++ #address-cells = <3>; ++ #size-cells = <2>; ++ device_type = "pci"; ++ num-lanes = <4>; ++ bus-range = <0x0 0xff>; ++ ranges = <0x81000000 0x0 0x00000000 0x40 0x00010000 0x0 0x00010000 /* downstream I/O */ ++ 0x82000000 0x0 0x40000000 0x40 0x40000000 0x0 0x40000000>; /* non-prefetchable memory */ ++ msi-parent = <&msi2>; ++ #interrupt-cells = <1>; ++ interrupt-map-mask = <0 0 0 7>; ++ interrupt-map = <0000 0 0 1 &gic 0 110 0x4>, ++ <0000 0 0 2 &gic 0 111 0x4>, ++ <0000 0 0 3 &gic 0 112 0x4>, ++ <0000 0 0 4 &gic 0 113 0x4>; ++ }; ++ }; ++ reserved-memory { ++ #address-cells = <2>; ++ #size-cells = <2>; ++ ranges; ++ ++ pfe_reserved: packetbuffer@83400000 { ++ reg = <0 0x83400000 0 0xc00000>; ++ }; ++ }; ++ ++ pfe: pfe@04000000 { ++ compatible = "fsl,pfe"; ++ ranges = <0x0 0x00 0x04000000 0xc00000 ++ 0x1 0x00 0x83400000 0xc00000>; ++ reg = <0x0 0x90500000 0x0 0x10000>, /* APB 64K */ ++ <0x0 0x04000000 0x0 0xc00000>, /* AXI 16M */ ++ <0x0 0x83400000 0x0 0xc00000>, /* PFE DDR 12M */ ++ <0x0 0x10000000 0x0 0x2000>; /* OCRAM 8K */ ++ fsl,pfe-num-interfaces = < 0x2 >; ++ interrupts = <0 172 0x4>; ++ #interrupt-names = "hifirq"; ++ memory-region = <&pfe_reserved>; ++ fsl,pfe-scfg = <&scfg 0>; ++ }; ++ ++}; diff --git a/target/linux/layerscape/patches-4.4/3117-armv8-aarch32-Run-32-bit-Linux-for-LayerScape-SoCs.patch b/target/linux/layerscape/patches-4.4/3117-armv8-aarch32-Run-32-bit-Linux-for-LayerScape-SoCs.patch new file mode 100644 index 0000000..9b580b4 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/3117-armv8-aarch32-Run-32-bit-Linux-for-LayerScape-SoCs.patch @@ -0,0 +1,49 @@ +From f6dcf8936845ea95eba432ee725cec761032fe2a Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 10:12:29 +0800 +Subject: [PATCH 117/124] armv8: aarch32: Run 32-bit Linux for LayerScape SoCs + +This patch adds AArch32 execution state support for LayerScape SoCs. + +Signed-off-by: Ebony Zhu <ebony.zhu@nxp.com> +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/mach-imx/Makefile | 2 ++ + arch/arm/mach-imx/mach-layerscape.c | 22 ++++++++++++++++++++++ + 2 files changed, 24 insertions(+) + create mode 100644 arch/arm/mach-imx/mach-layerscape.c + +--- a/arch/arm/mach-imx/Makefile ++++ b/arch/arm/mach-imx/Makefile +@@ -103,4 +103,6 @@ obj-$(CONFIG_SOC_LS1021A) += mach-ls1021 + + obj-$(CONFIG_ARCH_LAYERSCAPE) += mach-ls1043a.o + ++obj-$(CONFIG_ARCH_LAYERSCAPE) += mach-layerscape.o ++ + obj-y += devices/ +--- /dev/null ++++ b/arch/arm/mach-imx/mach-layerscape.c +@@ -0,0 +1,22 @@ ++/* ++ * Copyright 2015-2016 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ */ ++ ++#include <asm/mach/arch.h> ++ ++#include "common.h" ++ ++static const char * const layerscape_dt_compat[] __initconst = { ++ "fsl,ls1043a", ++ "fsl,ls1012a", ++ NULL, ++}; ++ ++DT_MACHINE_START(LAYERSCAPE, "Freescale LAYERSCAPE") ++ .dt_compat = layerscape_dt_compat, ++MACHINE_END diff --git a/target/linux/layerscape/patches-4.4/3118-armv8-aarch32-Add-KVM-support-for-AArch32-on-ARMv8.patch b/target/linux/layerscape/patches-4.4/3118-armv8-aarch32-Add-KVM-support-for-AArch32-on-ARMv8.patch new file mode 100644 index 0000000..32ebc36 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/3118-armv8-aarch32-Add-KVM-support-for-AArch32-on-ARMv8.patch @@ -0,0 +1,34 @@ +From 44914d3b7baff0f96c3584e4d46aa992e402ef4f Mon Sep 17 00:00:00 2001 +From: Ying Zhang <ying.zhang22455@nxp.com> +Date: Thu, 29 Sep 2016 11:16:16 +0800 +Subject: [PATCH 118/124] armv8: aarch32: Add KVM support for AArch32 on ARMv8 + +This patch is to add KVM support for AArch32 execution state on ARMv8. + +Signed-off-by: Alison Wang <alison.wang@nxp.com> +--- + arch/arm/include/asm/cputype.h | 2 ++ + arch/arm/kvm/guest.c | 1 + + 2 files changed, 3 insertions(+) + +--- a/arch/arm/include/asm/cputype.h ++++ b/arch/arm/include/asm/cputype.h +@@ -76,6 +76,8 @@ + #define ARM_CPU_PART_CORTEX_A15 0x4100c0f0 + #define ARM_CPU_PART_MASK 0xff00fff0 + ++#define ARM_CPU_PART_CORTEX_AARCH32_A53 0x4100d030 ++ + #define ARM_CPU_XSCALE_ARCH_MASK 0xe000 + #define ARM_CPU_XSCALE_ARCH_V1 0x2000 + #define ARM_CPU_XSCALE_ARCH_V2 0x4000 +--- a/arch/arm/kvm/guest.c ++++ b/arch/arm/kvm/guest.c +@@ -247,6 +247,7 @@ int __attribute_const__ kvm_target_cpu(v + { + switch (read_cpuid_part()) { + case ARM_CPU_PART_CORTEX_A7: ++ case ARM_CPU_PART_CORTEX_AARCH32_A53: + return KVM_ARM_TARGET_CORTEX_A7; + case ARM_CPU_PART_CORTEX_A15: + return KVM_ARM_TARGET_CORTEX_A15; diff --git a/target/linux/layerscape/patches-4.4/7072-LS1012-Add-PPFE-driver-in-Linux.patch b/target/linux/layerscape/patches-4.4/7072-LS1012-Add-PPFE-driver-in-Linux.patch new file mode 100644 index 0000000..21339e8 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/7072-LS1012-Add-PPFE-driver-in-Linux.patch @@ -0,0 +1,15167 @@ +From 0157efe2fbe2fe56c34727d326cd74284c06cbd5 Mon Sep 17 00:00:00 2001 +From: Bhaskar Upadhaya <Bhaskar.Upadhaya@freescale.com> +Date: Wed, 24 Aug 2016 10:51:21 +0800 +Subject: [PATCH 072/113] LS1012: Add PPFE driver in Linux + +commit 7584b690d4c8e4e435c2e6abcdb38d6595a0c302 +[context adjustment] +[don't apply fsl-ls1012a-rdb.dts and fsl-ls1012a.dtsi] +[Let PPFE driver can be selectd as a module] + +Signed-off-by: Bhaskar Upadhaya <Bhaskar.Upadhaya@freescale.com> +Integrated-by: Zhao Qiang <qiang.zhao@nxp.com> +Integrated-by: Yutang Jiang <yutang.jiang@nxp.com> +--- + drivers/staging/Kconfig | 2 + + drivers/staging/Makefile | 1 + + drivers/staging/fsl_ppfe/Kconfig | 5 + + drivers/staging/fsl_ppfe/Makefile | 44 + + drivers/staging/fsl_ppfe/config.h | 8 + + drivers/staging/fsl_ppfe/control_link.lds | 32 + + drivers/staging/fsl_ppfe/include/pfe/cbus.h | 88 + + drivers/staging/fsl_ppfe/include/pfe/cbus/bmu.h | 55 + + .../staging/fsl_ppfe/include/pfe/cbus/class_csr.h | 242 ++ + drivers/staging/fsl_ppfe/include/pfe/cbus/emac.h | 243 ++ + .../staging/fsl_ppfe/include/pfe/cbus/emac_mtip.h | 250 ++ + drivers/staging/fsl_ppfe/include/pfe/cbus/gpi.h | 78 + + drivers/staging/fsl_ppfe/include/pfe/cbus/gpt.h | 29 + + drivers/staging/fsl_ppfe/include/pfe/cbus/hif.h | 96 + + .../staging/fsl_ppfe/include/pfe/cbus/hif_nocpy.h | 51 + + .../staging/fsl_ppfe/include/pfe/cbus/tmu_csr.h | 128 + + .../staging/fsl_ppfe/include/pfe/cbus/util_csr.h | 61 + + drivers/staging/fsl_ppfe/include/pfe/class.h | 133 + + drivers/staging/fsl_ppfe/include/pfe/class/ccu.h | 28 + + drivers/staging/fsl_ppfe/include/pfe/class/efet.h | 44 + + .../staging/fsl_ppfe/include/pfe/class/mac_hash.h | 55 + + drivers/staging/fsl_ppfe/include/pfe/class/perg.h | 39 + + .../staging/fsl_ppfe/include/pfe/class/vlan_hash.h | 46 + + drivers/staging/fsl_ppfe/include/pfe/gpt.h | 44 + + drivers/staging/fsl_ppfe/include/pfe/pe.h | 626 +++++ + drivers/staging/fsl_ppfe/include/pfe/pfe.h | 444 +++ + drivers/staging/fsl_ppfe/include/pfe/tmu.h | 68 + + .../staging/fsl_ppfe/include/pfe/tmu/phy_queue.h | 56 + + drivers/staging/fsl_ppfe/include/pfe/tmu/sched.h | 72 + + drivers/staging/fsl_ppfe/include/pfe/tmu/shaper.h | 37 + + drivers/staging/fsl_ppfe/include/pfe/uart.h | 31 + + drivers/staging/fsl_ppfe/include/pfe/util.h | 49 + + drivers/staging/fsl_ppfe/include/pfe/util/eape.h | 57 + + drivers/staging/fsl_ppfe/include/pfe/util/efet.h | 119 + + drivers/staging/fsl_ppfe/include/pfe/util/inq.h | 28 + + drivers/staging/fsl_ppfe/pfe_ctrl.c | 363 +++ + drivers/staging/fsl_ppfe/pfe_ctrl.h | 111 + + drivers/staging/fsl_ppfe/pfe_ctrl_hal.c | 207 ++ + drivers/staging/fsl_ppfe/pfe_ctrl_hal.h | 129 + + drivers/staging/fsl_ppfe/pfe_debugfs.c | 109 + + drivers/staging/fsl_ppfe/pfe_debugfs.h | 8 + + drivers/staging/fsl_ppfe/pfe_eth.c | 2956 ++++++++++++++++++++ + drivers/staging/fsl_ppfe/pfe_eth.h | 384 +++ + drivers/staging/fsl_ppfe/pfe_firmware.c | 322 +++ + drivers/staging/fsl_ppfe/pfe_firmware.h | 41 + + drivers/staging/fsl_ppfe/pfe_hal.c | 2217 +++++++++++++++ + drivers/staging/fsl_ppfe/pfe_hif.c | 939 +++++++ + drivers/staging/fsl_ppfe/pfe_hif.h | 322 +++ + drivers/staging/fsl_ppfe/pfe_hif_lib.c | 658 +++++ + drivers/staging/fsl_ppfe/pfe_hif_lib.h | 219 ++ + drivers/staging/fsl_ppfe/pfe_hw.c | 188 ++ + drivers/staging/fsl_ppfe/pfe_hw.h | 32 + + drivers/staging/fsl_ppfe/pfe_ls1012a_platform.c | 341 +++ + drivers/staging/fsl_ppfe/pfe_mod.c | 140 + + drivers/staging/fsl_ppfe/pfe_mod.h | 163 ++ + drivers/staging/fsl_ppfe/pfe_perfmon.c | 175 ++ + drivers/staging/fsl_ppfe/pfe_perfmon.h | 41 + + drivers/staging/fsl_ppfe/pfe_platform.c | 358 +++ + drivers/staging/fsl_ppfe/pfe_sysfs.c | 855 ++++++ + drivers/staging/fsl_ppfe/pfe_sysfs.h | 34 + + drivers/staging/fsl_ppfe/platform.h | 25 + + include/linux/skbuff.h | 11 + + net/core/skbuff.c | 84 + + 63 files changed, 14821 insertions(+) + create mode 100644 drivers/staging/fsl_ppfe/Kconfig + create mode 100644 drivers/staging/fsl_ppfe/Makefile + create mode 100644 drivers/staging/fsl_ppfe/config.h + create mode 100644 drivers/staging/fsl_ppfe/control_link.lds + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/bmu.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/class_csr.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/emac.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/emac_mtip.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/gpi.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/gpt.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/hif.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/hif_nocpy.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/tmu_csr.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/cbus/util_csr.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/class.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/class/ccu.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/class/efet.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/class/mac_hash.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/class/perg.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/class/vlan_hash.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/gpt.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/pe.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/pfe.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/tmu.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/tmu/phy_queue.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/tmu/sched.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/tmu/shaper.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/uart.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/util.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/util/eape.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/util/efet.h + create mode 100644 drivers/staging/fsl_ppfe/include/pfe/util/inq.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_ctrl.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_ctrl.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_ctrl_hal.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_ctrl_hal.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_debugfs.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_debugfs.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_eth.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_eth.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_firmware.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_firmware.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_hal.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_hif.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_hif.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_hif_lib.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_hif_lib.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_hw.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_hw.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_ls1012a_platform.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_mod.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_mod.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_perfmon.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_perfmon.h + create mode 100644 drivers/staging/fsl_ppfe/pfe_platform.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_sysfs.c + create mode 100644 drivers/staging/fsl_ppfe/pfe_sysfs.h + create mode 100644 drivers/staging/fsl_ppfe/platform.h + +--- a/drivers/staging/Kconfig ++++ b/drivers/staging/Kconfig +@@ -112,4 +112,6 @@ source "drivers/staging/wilc1000/Kconfig + + source "drivers/staging/most/Kconfig" + ++source "drivers/staging/fsl_ppfe/Kconfig" ++ + endif # STAGING +--- a/drivers/staging/Makefile ++++ b/drivers/staging/Makefile +@@ -48,3 +48,4 @@ obj-$(CONFIG_FSL_MC_BUS) += fsl-mc/ + obj-$(CONFIG_FSL_DPA) += fsl_qbman/ + obj-$(CONFIG_WILC1000) += wilc1000/ + obj-$(CONFIG_MOST) += most/ ++obj-$(CONFIG_FSL_PPFE) += fsl_ppfe/ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/Kconfig +@@ -0,0 +1,5 @@ ++config FSL_PPFE ++ tristate "Freescale PPFE Driver" ++ default m ++ help ++ only compiled as module ! +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/Makefile +@@ -0,0 +1,44 @@ ++# ++# Copyright (C) 2007 Freescale Semiconductor, Inc. ++# ++# This program is free software; you can redistribute it and/or modify ++# it under the terms of the GNU General Public License as published by ++# the Free Software Foundation; either version 2 of the License, or ++# (at your option) any later version. ++# ++# This program is distributed in the hope that it will be useful, ++# but WITHOUT ANY WARRANTY; without even the implied warranty of ++# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++# GNU General Public License for more details. ++# ++# You should have received a copy of the GNU General Public License ++# along with this program; if not, write to the Free Software ++# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ ++ ++all: modules ++ ++modules clean: ++ make CROSS_COMPILE=$(CROSS_COMPILE) ARCH=$(ARCH) -C $(KERNELDIR) M=`pwd` $@ ++ ++EXTRA_CFLAGS += -I$(src)/include -I$(src) -DCOMCERTO_2000 -DCONFIG_PLATFORM_LS1012A -DGEMAC_MTIP -DCONFIG_UTIL_DISABLED ++ ++EXTRA_LDFLAGS += -T$(srctree)/$(src)/control_link.lds ++ ++#only compiled as module ! ++obj-$(CONFIG_FSL_PPFE) += pfe.o ++ ++pfe-y += pfe_mod.o \ ++ pfe_hw.o \ ++ pfe_firmware.o \ ++ pfe_ctrl.o \ ++ pfe_ctrl_hal.o \ ++ pfe_hif.o \ ++ pfe_hif_lib.o\ ++ pfe_eth.o \ ++ pfe_perfmon.o \ ++ pfe_sysfs.o \ ++ pfe_debugfs.o \ ++ pfe_ls1012a_platform.o \ ++ pfe_hal.o \ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/config.h +@@ -0,0 +1,8 @@ ++#ifndef _CONFIG_H_ ++#define _CONFIG_H_ ++#define CFG_WIFI_OFFLOAD (1 << 1) ++#define CFG_ICC (1 << 11) ++#define CFG_RTP (1 << 14) ++#define CFG_ELLIPTIC (1 << 15) ++#define CFG_ALL (0 | CFG_WIFI_OFFLOAD | CFG_ICC | CFG_RTP | CFG_ELLIPTIC ) ++#endif /* _CONFIG_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/control_link.lds +@@ -0,0 +1,32 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++SECTIONS ++{ ++ .class_dmem_sh : SUBALIGN(8) { ++ __class_dmem_sh = .; ++ *(SORT(.class_dmem_sh_*)) ++ } ++ ++ .tmu_dmem_sh : SUBALIGN(8) { ++ __tmu_dmem_sh = .; ++ *(SORT(.tmu_dmem_sh_*)) ++ } ++ ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus.h +@@ -0,0 +1,88 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _CBUS_H_ ++#define _CBUS_H_ ++ ++#define EMAC1_BASE_ADDR (CBUS_BASE_ADDR + 0x200000) ++#define EGPI1_BASE_ADDR (CBUS_BASE_ADDR + 0x210000) ++#define EMAC2_BASE_ADDR (CBUS_BASE_ADDR + 0x220000) ++#define EGPI2_BASE_ADDR (CBUS_BASE_ADDR + 0x230000) ++#define BMU1_BASE_ADDR (CBUS_BASE_ADDR + 0x240000) ++#define BMU2_BASE_ADDR (CBUS_BASE_ADDR + 0x250000) ++#define ARB_BASE_ADDR (CBUS_BASE_ADDR + 0x260000) /* FIXME not documented */ ++#define DDR_CONFIG_BASE_ADDR (CBUS_BASE_ADDR + 0x270000) /* FIXME not documented */ ++#define HIF_BASE_ADDR (CBUS_BASE_ADDR + 0x280000) ++#define HGPI_BASE_ADDR (CBUS_BASE_ADDR + 0x290000) ++#define LMEM_BASE_ADDR (CBUS_BASE_ADDR + 0x300000) ++#define LMEM_SIZE 0x10000 ++#define LMEM_END (LMEM_BASE_ADDR + LMEM_SIZE) ++#define TMU_CSR_BASE_ADDR (CBUS_BASE_ADDR + 0x310000) ++#define CLASS_CSR_BASE_ADDR (CBUS_BASE_ADDR + 0x320000) ++#if defined(CONFIG_PLATFORM_C2000) ++#define EMAC3_BASE_ADDR (CBUS_BASE_ADDR + 0x330000) ++#define EGPI3_BASE_ADDR (CBUS_BASE_ADDR + 0x340000) ++#endif ++#define HIF_NOCPY_BASE_ADDR (CBUS_BASE_ADDR + 0x350000) ++#define UTIL_CSR_BASE_ADDR (CBUS_BASE_ADDR + 0x360000) ++#define CBUS_GPT_BASE_ADDR (CBUS_BASE_ADDR + 0x370000) ++ ++#define IS_LMEM(addr, len) (((unsigned long)(addr) >= (unsigned long)LMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= (unsigned long)LMEM_END)) ++ ++/** ++* \defgroup XXX_MEM_ACCESS_ADDR PE memory access through CSR ++* XXX_MEM_ACCESS_ADDR register bit definitions. ++* @{ ++*/ ++#define PE_MEM_ACCESS_WRITE (1<<31) /**< Internal Memory Write. */ ++#define PE_MEM_ACCESS_IMEM (1<<15) ++#define PE_MEM_ACCESS_DMEM (1<<16) ++#define PE_MEM_ACCESS_BYTE_ENABLE(offset,size) (((((1 << (size)) - 1) << (4 - (offset) - (size))) & 0xf) << 24) /**< Byte Enables of the Internal memory access. These are interpred in BE */ ++// @} ++#if defined(CONFIG_PLATFORM_LS1012A) ++#include "cbus/emac_mtip.h" ++#else ++#include "cbus/emac.h" ++#endif //CONFIG_PLATFORM_LS1012A ++#include "cbus/gpi.h" ++#include "cbus/bmu.h" ++#include "cbus/hif.h" ++#include "cbus/tmu_csr.h" ++#include "cbus/class_csr.h" ++#include "cbus/hif_nocpy.h" ++#include "cbus/util_csr.h" ++#include "cbus/gpt.h" ++ ++ ++/* PFE cores states */ ++#define CORE_DISABLE 0x00000000 ++#define CORE_ENABLE 0x00000001 ++#define CORE_SW_RESET 0x00000002 ++ ++/* LMEM defines */ ++#define LMEM_HDR_SIZE 0x0010 ++#define LMEM_BUF_SIZE_LN2 0x7 ++#define LMEM_BUF_SIZE (1 << LMEM_BUF_SIZE_LN2) ++ ++/* DDR defines */ ++#define DDR_HDR_SIZE 0x0100 ++#define DDR_BUF_SIZE_LN2 0xb ++#define DDR_BUF_SIZE (1 << DDR_BUF_SIZE_LN2) ++ ++ ++#endif /* _CBUS_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/bmu.h +@@ -0,0 +1,55 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _BMU_H_ ++#define _BMU_H_ ++ ++#define BMU_VERSION 0x000 ++#define BMU_CTRL 0x004 ++#define BMU_UCAST_CONFIG 0x008 ++#define BMU_UCAST_BASE_ADDR 0x00c ++#define BMU_BUF_SIZE 0x010 ++#define BMU_BUF_CNT 0x014 ++#define BMU_THRES 0x018 ++#define BMU_INT_SRC 0x020 ++#define BMU_INT_ENABLE 0x024 ++#define BMU_ALLOC_CTRL 0x030 ++#define BMU_FREE_CTRL 0x034 ++#define BMU_FREE_ERR_ADDR 0x038 ++#define BMU_CURR_BUF_CNT 0x03c ++#define BMU_MCAST_CNT 0x040 ++#define BMU_MCAST_ALLOC_CTRL 0x044 ++#define BMU_REM_BUF_CNT 0x048 ++#define BMU_LOW_WATERMARK 0x050 ++#define BMU_HIGH_WATERMARK 0x054 ++#define BMU_INT_MEM_ACCESS 0x100 ++ ++typedef struct { ++ unsigned long baseaddr; ++ u32 count; ++ u32 size; ++} BMU_CFG; ++ ++ ++#define BMU1_BUF_SIZE LMEM_BUF_SIZE_LN2 ++#define BMU2_BUF_SIZE DDR_BUF_SIZE_LN2 ++ ++#define BMU2_MCAST_ALLOC_CTRL BMU2_BASE_ADDR + BMU_MCAST_ALLOC_CTRL ++ ++#endif /* _BMU_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/class_csr.h +@@ -0,0 +1,242 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _CLASS_CSR_H_ ++#define _CLASS_CSR_H_ ++ ++/** @file class_csr.h. ++ * class_csr - block containing all the classifier control and status register. Mapped on CBUS and accessible from all PE's and ARM. ++ */ ++ ++ ++#define CLASS_VERSION (CLASS_CSR_BASE_ADDR + 0x000) ++#define CLASS_TX_CTRL (CLASS_CSR_BASE_ADDR + 0x004) ++#define CLASS_INQ_PKTPTR (CLASS_CSR_BASE_ADDR + 0x010) ++#define CLASS_HDR_SIZE (CLASS_CSR_BASE_ADDR + 0x014) /**< (ddr_hdr_size[24:16], lmem_hdr_size[5:0]) */ ++#define CLASS_HDR_SIZE_LMEM(off) ((off) & 0x3f) /**< LMEM header size for the Classifier block.\ Data in the LMEM is written from this offset. */ ++#define CLASS_HDR_SIZE_DDR(off) (((off) & 0x1ff) << 16) /**< DDR header size for the Classifier block.\ Data in the DDR is written from this offset. */ ++ ++#define CLASS_PE0_QB_DM_ADDR0 (CLASS_CSR_BASE_ADDR + 0x020) /**< DMEM address of first [15:0] and second [31:16] buffers on QB side. */ ++#define CLASS_PE0_QB_DM_ADDR1 (CLASS_CSR_BASE_ADDR + 0x024) /**< DMEM address of third [15:0] and fourth [31:16] buffers on QB side. */ ++ ++#define CLASS_PE0_RO_DM_ADDR0 (CLASS_CSR_BASE_ADDR + 0x060) /**< DMEM address of first [15:0] and second [31:16] buffers on RO side. */ ++#define CLASS_PE0_RO_DM_ADDR1 (CLASS_CSR_BASE_ADDR + 0x064) /**< DMEM address of third [15:0] and fourth [31:16] buffers on RO side. */ ++ ++/** @name Class PE memory access. Allows external PE's and HOST to read/write PMEM/DMEM memory ranges for each classifier PE. ++ */ ++//@{ ++#define CLASS_MEM_ACCESS_ADDR (CLASS_CSR_BASE_ADDR + 0x100) /**< {sr_pe_mem_cmd[31], csr_pe_mem_wren[27:24], csr_pe_mem_addr[23:0]}, See \ref XXX_MEM_ACCESS_ADDR for details. */ ++#define CLASS_MEM_ACCESS_WDATA (CLASS_CSR_BASE_ADDR + 0x104) /**< Internal Memory Access Write Data [31:0] */ ++#define CLASS_MEM_ACCESS_RDATA (CLASS_CSR_BASE_ADDR + 0x108) /**< Internal Memory Access Read Data [31:0] */ ++//@} ++#define CLASS_TM_INQ_ADDR (CLASS_CSR_BASE_ADDR + 0x114) ++#define CLASS_PE_STATUS (CLASS_CSR_BASE_ADDR + 0x118) ++ ++#define CLASS_PHY1_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x11c) ++#define CLASS_PHY1_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x120) ++#define CLASS_PHY1_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x124) ++#define CLASS_PHY1_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x128) ++#define CLASS_PHY1_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x12c) ++#define CLASS_PHY1_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x130) ++#define CLASS_PHY1_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x134) ++#define CLASS_PHY1_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x138) ++#define CLASS_PHY1_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x13c) ++#define CLASS_PHY1_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x140) ++#define CLASS_PHY2_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x144) ++#define CLASS_PHY2_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x148) ++#define CLASS_PHY2_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x14c) ++#define CLASS_PHY2_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x150) ++#define CLASS_PHY2_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x154) ++#define CLASS_PHY2_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x158) ++#define CLASS_PHY2_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x15c) ++#define CLASS_PHY2_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x160) ++#define CLASS_PHY2_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x164) ++#define CLASS_PHY2_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x168) ++#define CLASS_PHY3_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x16c) ++#define CLASS_PHY3_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x170) ++#define CLASS_PHY3_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x174) ++#define CLASS_PHY3_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x178) ++#define CLASS_PHY3_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x17c) ++#define CLASS_PHY3_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x180) ++#define CLASS_PHY3_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x184) ++#define CLASS_PHY3_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x188) ++#define CLASS_PHY3_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x18c) ++#define CLASS_PHY3_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x190) ++#define CLASS_PHY1_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x194) ++#define CLASS_PHY1_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x198) ++#define CLASS_PHY1_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x19c) ++#define CLASS_PHY1_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1a0) ++#define CLASS_PHY2_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1a4) ++#define CLASS_PHY2_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1a8) ++#define CLASS_PHY2_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x1ac) ++#define CLASS_PHY2_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1b0) ++#define CLASS_PHY3_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1b4) ++#define CLASS_PHY3_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1b8) ++#define CLASS_PHY3_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x1bc) ++#define CLASS_PHY3_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1c0) ++#define CLASS_PHY4_ICMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1c4) ++#define CLASS_PHY4_IGMP_PKTS (CLASS_CSR_BASE_ADDR + 0x1c8) ++#define CLASS_PHY4_TCP_PKTS (CLASS_CSR_BASE_ADDR + 0x1cc) ++#define CLASS_PHY4_UDP_PKTS (CLASS_CSR_BASE_ADDR + 0x1d0) ++#define CLASS_PHY4_RX_PKTS (CLASS_CSR_BASE_ADDR + 0x1d4) ++#define CLASS_PHY4_TX_PKTS (CLASS_CSR_BASE_ADDR + 0x1d8) ++#define CLASS_PHY4_LP_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x1dc) ++#define CLASS_PHY4_INTF_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x1e0) ++#define CLASS_PHY4_INTF_MATCH_PKTS (CLASS_CSR_BASE_ADDR + 0x1e4) ++#define CLASS_PHY4_L3_FAIL_PKTS (CLASS_CSR_BASE_ADDR + 0x1e8) ++#define CLASS_PHY4_V4_PKTS (CLASS_CSR_BASE_ADDR + 0x1ec) ++#define CLASS_PHY4_V6_PKTS (CLASS_CSR_BASE_ADDR + 0x1f0) ++#define CLASS_PHY4_CHKSUM_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x1f4) ++#define CLASS_PHY4_TTL_ERR_PKTS (CLASS_CSR_BASE_ADDR + 0x1f8) ++ ++#define CLASS_PE_SYS_CLK_RATIO (CLASS_CSR_BASE_ADDR + 0x200) ++#define CLASS_AFULL_THRES (CLASS_CSR_BASE_ADDR + 0x204) ++#define CLASS_GAP_BETWEEN_READS (CLASS_CSR_BASE_ADDR + 0x208) ++#define CLASS_MAX_BUF_CNT (CLASS_CSR_BASE_ADDR + 0x20c) ++#define CLASS_TSQ_FIFO_THRES (CLASS_CSR_BASE_ADDR + 0x210) ++#define CLASS_TSQ_MAX_CNT (CLASS_CSR_BASE_ADDR + 0x214) ++#define CLASS_IRAM_DATA_0 (CLASS_CSR_BASE_ADDR + 0x218) ++#define CLASS_IRAM_DATA_1 (CLASS_CSR_BASE_ADDR + 0x21c) ++#define CLASS_IRAM_DATA_2 (CLASS_CSR_BASE_ADDR + 0x220) ++#define CLASS_IRAM_DATA_3 (CLASS_CSR_BASE_ADDR + 0x224) ++ ++#define CLASS_BUS_ACCESS_ADDR (CLASS_CSR_BASE_ADDR + 0x228) ++ ++#define CLASS_BUS_ACCESS_WDATA (CLASS_CSR_BASE_ADDR + 0x22c) ++#define CLASS_BUS_ACCESS_RDATA (CLASS_CSR_BASE_ADDR + 0x230) ++ ++#define CLASS_ROUTE_HASH_ENTRY_SIZE (CLASS_CSR_BASE_ADDR + 0x234) /**< (route_entry_size[9:0], route_hash_size[23:16] (this is actually ln2(size))) */ ++#define CLASS_ROUTE_ENTRY_SIZE(size) ((size) & 0x1ff) ++#define CLASS_ROUTE_HASH_SIZE(hash_bits) (((hash_bits) & 0xff) << 16) ++ ++#define CLASS_ROUTE_TABLE_BASE (CLASS_CSR_BASE_ADDR + 0x238) ++ ++#define CLASS_ROUTE_MULTI (CLASS_CSR_BASE_ADDR + 0x23c) ++#define CLASS_SMEM_OFFSET (CLASS_CSR_BASE_ADDR + 0x240) ++#define CLASS_LMEM_BUF_SIZE (CLASS_CSR_BASE_ADDR + 0x244) ++#define CLASS_VLAN_ID (CLASS_CSR_BASE_ADDR + 0x248) ++#define CLASS_BMU1_BUF_FREE (CLASS_CSR_BASE_ADDR + 0x24c) ++#define CLASS_USE_TMU_INQ (CLASS_CSR_BASE_ADDR + 0x250) ++#define CLASS_VLAN_ID1 (CLASS_CSR_BASE_ADDR + 0x254) ++ ++#define CLASS_BUS_ACCESS_BASE (CLASS_CSR_BASE_ADDR + 0x258) ++#define CLASS_BUS_ACCESS_BASE_MASK (0xFF000000) //bit 31:24 of PE peripheral address are stored in CLASS_BUS_ACCESS_BASE ++ ++#define CLASS_HIF_PARSE (CLASS_CSR_BASE_ADDR + 0x25c) ++ ++#define CLASS_HOST_PE0_GP (CLASS_CSR_BASE_ADDR + 0x260) ++#define CLASS_PE0_GP (CLASS_CSR_BASE_ADDR + 0x264) ++#define CLASS_HOST_PE1_GP (CLASS_CSR_BASE_ADDR + 0x268) ++#define CLASS_PE1_GP (CLASS_CSR_BASE_ADDR + 0x26c) ++#define CLASS_HOST_PE2_GP (CLASS_CSR_BASE_ADDR + 0x270) ++#define CLASS_PE2_GP (CLASS_CSR_BASE_ADDR + 0x274) ++#define CLASS_HOST_PE3_GP (CLASS_CSR_BASE_ADDR + 0x278) ++#define CLASS_PE3_GP (CLASS_CSR_BASE_ADDR + 0x27c) ++#define CLASS_HOST_PE4_GP (CLASS_CSR_BASE_ADDR + 0x280) ++#define CLASS_PE4_GP (CLASS_CSR_BASE_ADDR + 0x284) ++#define CLASS_HOST_PE5_GP (CLASS_CSR_BASE_ADDR + 0x288) ++#define CLASS_PE5_GP (CLASS_CSR_BASE_ADDR + 0x28c) ++ ++#define CLASS_PE_INT_SRC (CLASS_CSR_BASE_ADDR + 0x290) ++#define CLASS_PE_INT_ENABLE (CLASS_CSR_BASE_ADDR + 0x294) ++ ++#define CLASS_TPID0_TPID1 (CLASS_CSR_BASE_ADDR + 0x298) ++#define CLASS_TPID2 (CLASS_CSR_BASE_ADDR + 0x29c) ++ ++#define CLASS_L4_CHKSUM_ADDR (CLASS_CSR_BASE_ADDR + 0x2a0) ++ ++#define CLASS_PE0_DEBUG (CLASS_CSR_BASE_ADDR + 0x2a4) ++#define CLASS_PE1_DEBUG (CLASS_CSR_BASE_ADDR + 0x2a8) ++#define CLASS_PE2_DEBUG (CLASS_CSR_BASE_ADDR + 0x2ac) ++#define CLASS_PE3_DEBUG (CLASS_CSR_BASE_ADDR + 0x2b0) ++#define CLASS_PE4_DEBUG (CLASS_CSR_BASE_ADDR + 0x2b4) ++#define CLASS_PE5_DEBUG (CLASS_CSR_BASE_ADDR + 0x2b8) ++ ++#define CLASS_STATE (CLASS_CSR_BASE_ADDR + 0x2bc) ++ ++/* CLASS defines */ ++#define CLASS_PBUF_SIZE 0x100 /* Fixed by hardware */ ++#define CLASS_PBUF_HEADER_OFFSET 0x80 /* Can be configured */ ++ ++#define CLASS_PBUF0_BASE_ADDR 0x000 /* Can be configured */ ++#define CLASS_PBUF1_BASE_ADDR (CLASS_PBUF0_BASE_ADDR + CLASS_PBUF_SIZE) /* Can be configured */ ++#define CLASS_PBUF2_BASE_ADDR (CLASS_PBUF1_BASE_ADDR + CLASS_PBUF_SIZE) /* Can be configured */ ++#define CLASS_PBUF3_BASE_ADDR (CLASS_PBUF2_BASE_ADDR + CLASS_PBUF_SIZE) /* Can be configured */ ++ ++#define CLASS_PBUF0_HEADER_BASE_ADDR (CLASS_PBUF0_BASE_ADDR + CLASS_PBUF_HEADER_OFFSET) ++#define CLASS_PBUF1_HEADER_BASE_ADDR (CLASS_PBUF1_BASE_ADDR + CLASS_PBUF_HEADER_OFFSET) ++#define CLASS_PBUF2_HEADER_BASE_ADDR (CLASS_PBUF2_BASE_ADDR + CLASS_PBUF_HEADER_OFFSET) ++#define CLASS_PBUF3_HEADER_BASE_ADDR (CLASS_PBUF3_BASE_ADDR + CLASS_PBUF_HEADER_OFFSET) ++ ++#define CLASS_PE0_RO_DM_ADDR0_VAL ((CLASS_PBUF1_BASE_ADDR << 16) | CLASS_PBUF0_BASE_ADDR) ++#define CLASS_PE0_RO_DM_ADDR1_VAL ((CLASS_PBUF3_BASE_ADDR << 16) | CLASS_PBUF2_BASE_ADDR) ++ ++#define CLASS_PE0_QB_DM_ADDR0_VAL ((CLASS_PBUF1_HEADER_BASE_ADDR << 16) | CLASS_PBUF0_HEADER_BASE_ADDR) ++#define CLASS_PE0_QB_DM_ADDR1_VAL ((CLASS_PBUF3_HEADER_BASE_ADDR << 16) | CLASS_PBUF2_HEADER_BASE_ADDR) ++ ++#define CLASS_ROUTE_SIZE 128 ++#define CLASS_MAX_ROUTE_SIZE 256 ++#define CLASS_ROUTE_HASH_BITS 20 ++#define CLASS_ROUTE_HASH_MASK ((1 << CLASS_ROUTE_HASH_BITS) - 1) ++ ++#define CLASS_ROUTE0_BASE_ADDR 0x400 /* Can be configured */ ++#define CLASS_ROUTE1_BASE_ADDR (CLASS_ROUTE0_BASE_ADDR + CLASS_ROUTE_SIZE) /* Can be configured */ ++#define CLASS_ROUTE2_BASE_ADDR (CLASS_ROUTE1_BASE_ADDR + CLASS_ROUTE_SIZE) /* Can be configured */ ++#define CLASS_ROUTE3_BASE_ADDR (CLASS_ROUTE2_BASE_ADDR + CLASS_ROUTE_SIZE) /* Can be configured */ ++ ++#define CLASS_SA_SIZE 128 ++#define CLASS_IPSEC_SA0_BASE_ADDR 0x600 ++#define CLASS_IPSEC_SA1_BASE_ADDR (CLASS_IPSEC_SA0_BASE_ADDR + CLASS_SA_SIZE) /* not used */ ++#define CLASS_IPSEC_SA2_BASE_ADDR (CLASS_IPSEC_SA1_BASE_ADDR + CLASS_SA_SIZE) /* not used */ ++#define CLASS_IPSEC_SA3_BASE_ADDR (CLASS_IPSEC_SA2_BASE_ADDR + CLASS_SA_SIZE) /* not used */ ++ ++/* generic purpose free dmem buffer, last portion of 2K dmem pbuf */ ++#define CLASS_GP_DMEM_BUF_SIZE (2048 - (CLASS_PBUF_SIZE*4) - (CLASS_ROUTE_SIZE*4) - (CLASS_SA_SIZE)) ++#define CLASS_GP_DMEM_BUF ((void *)(CLASS_IPSEC_SA0_BASE_ADDR + CLASS_SA_SIZE)) ++ ++ ++#define TWO_LEVEL_ROUTE (1 << 0) ++#define PHYNO_IN_HASH (1 << 1) ++#define HW_ROUTE_FETCH (1 << 3) ++#define HW_BRIDGE_FETCH (1 << 5) ++#define IP_ALIGNED (1 << 6) ++#define ARC_HIT_CHECK_EN (1 << 7) ++#define CLASS_TOE (1 << 11) ++#define HASH_NORMAL (0 << 12) ++#define HASH_CRC_PORT (1 << 12) ++#define HASH_CRC_IP (2 << 12) ++#define HASH_CRC_PORT_IP (3 << 12) ++#define QB2BUS_LE (1 << 15) ++ ++#define TCP_CHKSUM_DROP (1 << 0) ++#define UDP_CHKSUM_DROP (1 << 1) ++#define IPV4_CHKSUM_DROP (1 << 9) ++ ++/*CLASS_HIF_PARSE bits*/ ++#define HIF_PKT_CLASS_EN (1 << 0) ++#define HIF_PKT_OFFSET(ofst) ((ofst&0xF) << 1) ++ ++typedef struct { ++ u32 toe_mode; ++ unsigned long route_table_baseaddr; ++ u32 route_table_hash_bits; ++ u32 pe_sys_clk_ratio; ++ u32 resume; ++} CLASS_CFG; ++ ++#endif /* _CLASS_CSR_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/emac.h +@@ -0,0 +1,243 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _EMAC_H_ ++#define _EMAC_H_ ++ ++#define EMAC_NETWORK_CONTROL 0x000 ++#define EMAC_NETWORK_CONFIG 0x004 ++#define EMAC_NETWORK_STATUS 0x008 ++#define EMAC_DMA_CONFIG 0x010 ++ ++#define EMAC_PHY_MANAGEMENT 0x034 ++ ++#define EMAC_HASH_BOT 0x080 ++#define EMAC_HASH_TOP 0x084 ++ ++#define EMAC_SPEC1_ADD_BOT 0x088 ++#define EMAC_SPEC1_ADD_TOP 0x08c ++#define EMAC_SPEC2_ADD_BOT 0x090 ++#define EMAC_SPEC2_ADD_TOP 0x094 ++#define EMAC_SPEC3_ADD_BOT 0x098 ++#define EMAC_SPEC3_ADD_TOP 0x09c ++#define EMAC_SPEC4_ADD_BOT 0x0a0 ++#define EMAC_SPEC4_ADD_TOP 0x0a4 ++#define EMAC_WOL 0x0b8 ++ ++#define EMAC_STACKED_VLAN_REG 0x0c0 ++ ++#define EMAC_SPEC1_ADD_MASK_BOT 0x0c8 ++#define EMAC_SPEC1_ADD_MASK_TOP 0x0cc ++ ++#define EMAC_RMON_BASE_OFST 0x100 ++ ++#define EMAC_SPEC5_ADD_BOT 0x300 ++#define EMAC_SPEC5_ADD_TOP 0x304 ++#define EMAC_SPEC6_ADD_BOT 0x308 ++#define EMAC_SPEC6_ADD_TOP 0x30c ++#define EMAC_SPEC7_ADD_BOT 0x310 ++#define EMAC_SPEC7_ADD_TOP 0x314 ++#define EMAC_SPEC8_ADD_BOT 0x318 ++#define EMAC_SPEC8_ADD_TOP 0x31c ++#define EMAC_SPEC9_ADD_BOT 0x320 ++#define EMAC_SPEC9_ADD_TOP 0x324 ++#define EMAC_SPEC10_ADD_BOT 0x328 ++#define EMAC_SPEC10_ADD_TOP 0x32c ++#define EMAC_SPEC11_ADD_BOT 0x330 ++#define EMAC_SPEC11_ADD_TOP 0x334 ++#define EMAC_SPEC12_ADD_BOT 0x338 ++#define EMAC_SPEC12_ADD_TOP 0x33c ++#define EMAC_SPEC13_ADD_BOT 0x340 ++#define EMAC_SPEC13_ADD_TOP 0x344 ++#define EMAC_SPEC14_ADD_BOT 0x348 ++#define EMAC_SPEC14_ADD_TOP 0x34c ++#define EMAC_SPEC15_ADD_BOT 0x350 ++#define EMAC_SPEC15_ADD_TOP 0x354 ++#define EMAC_SPEC16_ADD_BOT 0x358 ++#define EMAC_SPEC16_ADD_TOP 0x35c ++#define EMAC_SPEC17_ADD_BOT 0x360 ++#define EMAC_SPEC17_ADD_TOP 0x364 ++#define EMAC_SPEC18_ADD_BOT 0x368 ++#define EMAC_SPEC18_ADD_TOP 0x36c ++#define EMAC_SPEC19_ADD_BOT 0x370 ++#define EMAC_SPEC19_ADD_TOP 0x374 ++#define EMAC_SPEC20_ADD_BOT 0x378 ++#define EMAC_SPEC20_ADD_TOP 0x37c ++#define EMAC_SPEC21_ADD_BOT 0x380 ++#define EMAC_SPEC21_ADD_TOP 0x384 ++#define EMAC_SPEC22_ADD_BOT 0x388 ++#define EMAC_SPEC22_ADD_TOP 0x38c ++#define EMAC_SPEC23_ADD_BOT 0x390 ++#define EMAC_SPEC23_ADD_TOP 0x394 ++#define EMAC_SPEC24_ADD_BOT 0x398 ++#define EMAC_SPEC24_ADD_TOP 0x39c ++#define EMAC_SPEC25_ADD_BOT 0x3a0 ++#define EMAC_SPEC25_ADD_TOP 0x3a4 ++#define EMAC_SPEC26_ADD_BOT 0x3a8 ++#define EMAC_SPEC26_ADD_TOP 0x3ac ++#define EMAC_SPEC27_ADD_BOT 0x3b0 ++#define EMAC_SPEC27_ADD_TOP 0x3b4 ++#define EMAC_SPEC28_ADD_BOT 0x3b8 ++#define EMAC_SPEC28_ADD_TOP 0x3bc ++#define EMAC_SPEC29_ADD_BOT 0x3c0 ++#define EMAC_SPEC29_ADD_TOP 0x3c4 ++#define EMAC_SPEC30_ADD_BOT 0x3c8 ++#define EMAC_SPEC30_ADD_TOP 0x3cc ++#define EMAC_SPEC31_ADD_BOT 0x3d0 ++#define EMAC_SPEC31_ADD_TOP 0x3d4 ++#define EMAC_SPEC32_ADD_BOT 0x3d8 ++#define EMAC_SPEC32_ADD_TOP 0x3dc ++ ++#define EMAC_SPEC_ADDR_MAX 32 ++ ++#define EMAC_CONTROL 0x7a0 ++ ++/* GEMAC definitions and settings */ ++ ++#define EMAC_PORT_0 0 ++#define EMAC_PORT_1 1 ++#define EMAC_PORT_2 2 ++ ++/* The possible operating speeds of the MAC, currently supporting 10, 100 and ++ * 1000Mb modes. ++ */ ++typedef enum {SPEED_10M, SPEED_100M, SPEED_1000M, SPEED_1000M_PCS} MAC_SPEED; ++ ++#define GMII 1 ++#define MII 2 ++#define RMII 3 ++#define RGMII 4 ++#define SGMII 5 ++ ++#define DUP_HALF 0x00 ++#define DUP_FULL 0x01 ++ ++/* EMAC_NETWORK_CONTROL bits definition */ ++#define EMAC_LB_PHY (1 << 0) ++#define EMAC_LB_MAC (1 << 1) ++#define EMAC_RX_ENABLE (1 << 2) ++#define EMAC_TX_ENABLE (1 << 3) ++#define EMAC_MDIO_EN (1 << 4) /* Enable MDIO port */ ++ ++/* WoL (Wake on Lan bit definition) */ ++#define EMAC_WOL_MAGIC (1 << 16) ++#define EMAC_WOL_ARP (1 << 17) ++#define EMAC_WOL_SPEC_ADDR (1 << 18) ++#define EMAC_WOL_MULTI (1 << 19) ++ ++/* EMAC_NETWORK_CONFIG bits definition */ ++#define EMAC_SPEED_100 (1 << 0) ++#define EMAC_HALF_DUP (0 << 1) ++#define EMAC_FULL_DUP (1 << 1) ++#define EMAC_DUPLEX_MASK (1 << 1) ++#define EMAC_ENABLE_JUMBO_FRAME (1 << 3) ++#define EMAC_ENABLE_COPY_ALL (1 << 4) ++#define EMAC_NO_BROADCAST (1 << 5) ++#define EMAC_ENABLE_MULTICAST (1 << 6) ++#define EMAC_ENABLE_UNICAST (1 << 7) ++#define EMAC_ENABLE_1536_RX (1 << 8) ++#define EMAC_SPEED_1000 (1 << 10) ++#define EMAC_PCS_ENABLE (1 << 11) ++#define EMAC_ENABLE_PAUSE_RX (1 << 13) ++#define EMAC_REMOVE_FCS (1 << 17) ++#define EMAC_ENABLE_CHKSUM_RX (1 << 24) ++#define EMAC_MDC_DIV_MASK (0x7 << 18) /* PCLK divisor for MDC */ ++#define EMAC_DATA_BUS_WIDTH_SHIFT 21 ++#define EMAC_DATA_BUS_WIDTH_MASK (0x3 << EMAC_DATA_BUS_WIDTH_SHIFT) ++#define EMAC_DATA_BUS_WIDTH_32 (0x00 << EMAC_DATA_BUS_WIDTH_SHIFT) ++#define EMAC_DATA_BUS_WIDTH_64 (0x01 << EMAC_DATA_BUS_WIDTH_SHIFT) ++#define EMAC_DATA_BUS_WIDTH_128 (0x10 << EMAC_DATA_BUS_WIDTH_SHIFT) ++#define EMAC_ENABLE_FCS_RX (1 << 26) ++#define EMAC_SGMII_MODE_ENABLE (1 << 27) ++ ++#define EMAC_SPEED_MASK (EMAC_SPEED_100 | EMAC_SPEED_1000) ++ ++/* EMAC_STACKED_VLAN_REG bits definition */ ++#define EMAC_ENABLE_STACKED_VLAN (1 << 31) ++ ++/* EMAC_CONTROL bits definition */ ++#define EMAC_TWO_BYTES_IP_ALIGN (1 << 0) // two bytes IP alignement ++ ++/* EMAC_NET_STATUS bits definition */ ++#define EMAC_PHY_IDLE (1<<2) /* PHY management is idle */ ++#define EMAC_MDIO_IN (1<<1) /* Status of mdio_in pin */ ++#define EMAC_LINK_STATUS (1<<0) /* Status of link pin */ ++ ++/* EMAC_DMA_CONFIG Bit definitions */ ++#define EMAC_ENABLE_CHKSUM_TX (1<<11) ++ ++//RMII enable – bit 1 / RGMII enable – bit 2 ++#define EMAC_RMII_MODE_ENABLE ((1 << 1) | (0 << 2)) ++#define EMAC_RMII_MODE_DISABLE (0 << 1) ++#define EMAC_RGMII_MODE_ENABLE ((0 << 1) | (1 << 2)) ++#define EMAC_RGMII_MODE_DISABLE (0 << 2) ++#define EMAC_MII_MODE_ENABLE (EMAC_RMII_MODE_DISABLE | EMAC_RGMII_MODE_DISABLE) ++#define EMAC_GMII_MODE_ENABLE (EMAC_RMII_MODE_DISABLE | EMAC_RGMII_MODE_DISABLE) ++#define EMAC_MODE_MASK (0x3 << 1) ++ ++/* Default configuration */ ++#define EMAC0_DEFAULT_DUPLEX_MODE FULLDUPLEX ++#define EMAC0_DEFAULT_EMAC_MODE RGMII ++#define EMAC0_DEFAULT_EMAC_SPEED SPEED_1000M ++ ++#define EMAC1_DEFAULT_DUPLEX_MODE FULLDUPLEX ++#define EMAC1_DEFAULT_EMAC_MODE RGMII ++#define EMAC1_DEFAULT_EMAC_SPEED SPEED_1000M ++ ++#define EMAC2_DEFAULT_DUPLEX_MODE FULLDUPLEX ++#define EMAC2_DEFAULT_EMAC_MODE RGMII ++#define EMAC2_DEFAULT_EMAC_SPEED SPEED_1000M ++ ++/* EMAC Hash size */ ++#define EMAC_HASH_REG_BITS 64 ++ ++/* The Address organisation for the MAC device. All addresses are split into ++ * two 32-bit register fields. The first one (bottom) is the lower 32-bits of ++ * the address and the other field are the high order bits - this may be 16-bits ++ * in the case of MAC addresses, or 32-bits for the hash address. ++ * In terms of memory storage, the first item (bottom) is assumed to be at a ++ * lower address location than 'top'. i.e. top should be at address location of ++ * 'bottom' + 4 bytes. ++ */ ++typedef struct { ++ u32 bottom; /* Lower 32-bits of address. */ ++ u32 top; /* Upper 32-bits of address. */ ++} MAC_ADDR; ++ ++ ++/* The following is the organisation of the address filters section of the MAC ++ * registers. The Cadence MAC contains four possible specific address match ++ * addresses, if an incoming frame corresponds to any one of these four ++ * addresses then the frame will be copied to memory. ++ * It is not necessary for all four of the address match registers to be ++ * programmed, this is application dependant. ++ */ ++typedef struct { ++ MAC_ADDR one; /* Specific address register 1. */ ++ MAC_ADDR two; /* Specific address register 2. */ ++ MAC_ADDR three; /* Specific address register 3. */ ++ MAC_ADDR four; /* Specific address register 4. */ ++} SPEC_ADDR; ++ ++typedef struct { ++ u32 mode; ++ u32 speed; ++ u32 duplex; ++} GEMAC_CFG; ++ ++#endif /* _EMAC_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/emac_mtip.h +@@ -0,0 +1,250 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _EMAC_H_ ++#define _EMAC_H_ ++ ++#define EMAC_IEVENT_REG 0x004 ++#define EMAC_IMASK_REG 0x008 ++#define EMAC_R_DES_ACTIVE_REG 0x010 ++#define EMAC_X_DES_ACTIVE_REG 0x014 ++#define EMAC_ECNTRL_REG 0x024 ++#define EMAC_MII_DATA_REG 0x040 ++#define EMAC_MII_CTRL_REG 0x044 ++#define EMAC_MIB_CTRL_STS_REG 0x064 ++#define EMAC_RCNTRL_REG 0x084 ++#define EMAC_TCNTRL_REG 0x0C4 ++#define EMAC_PHY_ADDR_LOW 0x0E4 ++#define EMAC_PHY_ADDR_HIGH 0x0E8 ++#define EMAC_GAUR 0x120 ++#define EMAC_GALR 0x124 ++#define EMAC_TFWR_STR_FWD 0x144 ++#define EMAC_RX_SECTIOM_FULL 0x190 ++#define EMAC_TX_SECTION_EMPTY 0x1A0 ++#define EMAC_TRUNC_FL 0x1B0 ++ ++#define RMON_T_DROP 0x200 /* Count of frames not cntd correctly */ ++#define RMON_T_PACKETS 0x204 /* RMON TX packet count */ ++#define RMON_T_BC_PKT 0x208 /* RMON TX broadcast pkts */ ++#define RMON_T_MC_PKT 0x20c /* RMON TX multicast pkts */ ++#define RMON_T_CRC_ALIGN 0x210 /* RMON TX pkts with CRC align err */ ++#define RMON_T_UNDERSIZE 0x214 /* RMON TX pkts < 64 bytes, good CRC */ ++#define RMON_T_OVERSIZE 0x218 /* RMON TX pkts > MAX_FL bytes good CRC */ ++#define RMON_T_FRAG 0x21c /* RMON TX pkts < 64 bytes, bad CRC */ ++#define RMON_T_JAB 0x220 /* RMON TX pkts > MAX_FL bytes, bad CRC */ ++#define RMON_T_COL 0x224 /* RMON TX collision count */ ++#define RMON_T_P64 0x228 /* RMON TX 64 byte pkts */ ++#define RMON_T_P65TO127 0x22c /* RMON TX 65 to 127 byte pkts */ ++#define RMON_T_P128TO255 0x230 /* RMON TX 128 to 255 byte pkts */ ++#define RMON_T_P256TO511 0x234 /* RMON TX 256 to 511 byte pkts */ ++#define RMON_T_P512TO1023 0x238 /* RMON TX 512 to 1023 byte pkts */ ++#define RMON_T_P1024TO2047 0x23c /* RMON TX 1024 to 2047 byte pkts */ ++#define RMON_T_P_GTE2048 0x240 /* RMON TX pkts > 2048 bytes */ ++#define RMON_T_OCTETS 0x244 /* RMON TX octets */ ++#define IEEE_T_DROP 0x248 /* Count of frames not counted crtly */ ++#define IEEE_T_FRAME_OK 0x24c /* Frames tx'd OK */ ++#define IEEE_T_1COL 0x250 /* Frames tx'd with single collision */ ++#define IEEE_T_MCOL 0x254 /* Frames tx'd with multiple collision */ ++#define IEEE_T_DEF 0x258 /* Frames tx'd after deferral delay */ ++#define IEEE_T_LCOL 0x25c /* Frames tx'd with late collision */ ++#define IEEE_T_EXCOL 0x260 /* Frames tx'd with excesv collisions */ ++#define IEEE_T_MACERR 0x264 /* Frames tx'd with TX FIFO underrun */ ++#define IEEE_T_CSERR 0x268 /* Frames tx'd with carrier sense err */ ++#define IEEE_T_SQE 0x26c /* Frames tx'd with SQE err */ ++#define IEEE_T_FDXFC 0x270 /* Flow control pause frames tx'd */ ++#define IEEE_T_OCTETS_OK 0x274 /* Octet count for frames tx'd w/o err */ ++#define RMON_R_PACKETS 0x284 /* RMON RX packet count */ ++#define RMON_R_BC_PKT 0x288 /* RMON RX broadcast pkts */ ++#define RMON_R_MC_PKT 0x28c /* RMON RX multicast pkts */ ++#define RMON_R_CRC_ALIGN 0x290 /* RMON RX pkts with CRC alignment err */ ++#define RMON_R_UNDERSIZE 0x294 /* RMON RX pkts < 64 bytes, good CRC */ ++#define RMON_R_OVERSIZE 0x298 /* RMON RX pkts > MAX_FL bytes good CRC */ ++#define RMON_R_FRAG 0x29c /* RMON RX pkts < 64 bytes, bad CRC */ ++#define RMON_R_JAB 0x2a0 /* RMON RX pkts > MAX_FL bytes, bad CRC */ ++#define RMON_R_RESVD_O 0x2a4 /* Reserved */ ++#define RMON_R_P64 0x2a8 /* RMON RX 64 byte pkts */ ++#define RMON_R_P65TO127 0x2ac /* RMON RX 65 to 127 byte pkts */ ++#define RMON_R_P128TO255 0x2b0 /* RMON RX 128 to 255 byte pkts */ ++#define RMON_R_P256TO511 0x2b4 /* RMON RX 256 to 511 byte pkts */ ++#define RMON_R_P512TO1023 0x2b8 /* RMON RX 512 to 1023 byte pkts */ ++#define RMON_R_P1024TO2047 0x2bc /* RMON RX 1024 to 2047 byte pkts */ ++#define RMON_R_P_GTE2048 0x2c0 /* RMON RX pkts > 2048 bytes */ ++#define RMON_R_OCTETS 0x2c4 /* RMON RX octets */ ++#define IEEE_R_DROP 0x2c8 /* Count frames not counted correctly */ ++#define IEEE_R_FRAME_OK 0x2cc /* Frames rx'd OK */ ++#define IEEE_R_CRC 0x2d0 /* Frames rx'd with CRC err */ ++#define IEEE_R_ALIGN 0x2d4 /* Frames rx'd with alignment err */ ++#define IEEE_R_MACERR 0x2d8 /* Receive FIFO overflow count */ ++#define IEEE_R_FDXFC 0x2dc /* Flow control pause frames rx'd */ ++#define IEEE_R_OCTETS_OK 0x2e0 /* Octet cnt for frames rx'd w/o err */ ++ ++#define EMAC_SMAC_0_0 0x500 /*Supplemental MAC Address 0 (RW).*/ ++#define EMAC_SMAC_0_1 0x504 /*Supplemental MAC Address 0 (RW).*/ ++ ++/* GEMAC definitions and settings */ ++ ++#define EMAC_PORT_0 0 ++#define EMAC_PORT_1 1 ++ ++/* GEMAC Bit definitions */ ++#define EMAC_IEVENT_HBERR 0x80000000 ++#define EMAC_IEVENT_BABR 0x40000000 ++#define EMAC_IEVENT_BABT 0x20000000 ++#define EMAC_IEVENT_GRA 0x10000000 ++#define EMAC_IEVENT_TXF 0x08000000 ++#define EMAC_IEVENT_TXB 0x04000000 ++#define EMAC_IEVENT_RXF 0x02000000 ++#define EMAC_IEVENT_RXB 0x01000000 ++#define EMAC_IEVENT_MII 0x00800000 ++#define EMAC_IEVENT_EBERR 0x00400000 ++#define EMAC_IEVENT_LC 0x00200000 ++#define EMAC_IEVENT_RL 0x00100000 ++#define EMAC_IEVENT_UN 0x00080000 ++ ++#define EMAC_IMASK_HBERR 0x80000000 ++#define EMAC_IMASK_BABR 0x40000000 ++#define EMAC_IMASKT_BABT 0x20000000 ++#define EMAC_IMASK_GRA 0x10000000 ++#define EMAC_IMASKT_TXF 0x08000000 ++#define EMAC_IMASK_TXB 0x04000000 ++#define EMAC_IMASKT_RXF 0x02000000 ++#define EMAC_IMASK_RXB 0x01000000 ++#define EMAC_IMASK_MII 0x00800000 ++#define EMAC_IMASK_EBERR 0x00400000 ++#define EMAC_IMASK_LC 0x00200000 ++#define EMAC_IMASKT_RL 0x00100000 ++#define EMAC_IMASK_UN 0x00080000 ++ ++#define EMAC_RCNTRL_MAX_FL_SHIFT 16 ++#define EMAC_RCNTRL_LOOP 0x00000001 ++#define EMAC_RCNTRL_DRT 0x00000002 ++#define EMAC_RCNTRL_MII_MODE 0x00000004 ++#define EMAC_RCNTRL_PROM 0x00000008 ++#define EMAC_RCNTRL_BC_REJ 0x00000010 ++#define EMAC_RCNTRL_FCE 0x00000020 ++#define EMAC_RCNTRL_RGMII 0x00000040 ++#define EMAC_RCNTRL_SGMII 0x00000080 ++#define EMAC_RCNTRL_RMII 0x00000100 ++#define EMAC_RCNTRL_RMII_10T 0x00000200 ++#define EMAC_RCNTRL_CRC_FWD 0x00004000 ++ ++#define EMAC_TCNTRL_GTS 0x00000001 ++#define EMAC_TCNTRL_HBC 0x00000002 ++#define EMAC_TCNTRL_FDEN 0x00000004 ++#define EMAC_TCNTRL_TFC_PAUSE 0x00000008 ++#define EMAC_TCNTRL_RFC_PAUSE 0x00000010 ++ ++#define EMAC_ECNTRL_RESET 0x00000001 /* reset the EMAC */ ++#define EMAC_ECNTRL_ETHER_EN 0x00000002 /* enable the EMAC */ ++#define EMAC_ECNTRL_SPEED 0x00000020 ++#define EMAC_ECNTRL_DBSWAP 0x00000100 ++ ++#define EMAC_X_WMRK_STRFWD 0x00000100 ++ ++#define EMAC_X_DES_ACTIVE_TDAR 0x01000000 ++#define EMAC_R_DES_ACTIVE_RDAR 0x01000000 ++ ++ ++ ++/* The possible operating speeds of the MAC, currently supporting 10, 100 and ++ * 1000Mb modes. ++ */ ++typedef enum {SPEED_10M, SPEED_100M, SPEED_1000M, SPEED_1000M_PCS} MAC_SPEED; ++ ++#define GMII 1 ++#define MII 2 ++#define RMII 3 ++#define RGMII 4 ++#define SGMII 5 ++ ++#define DUPLEX_HALF 0x00 ++#define DUPLEX_FULL 0x01 ++ ++ ++/* Default configuration */ ++#define EMAC0_DEFAULT_DUPLEX_MODE FULLDUPLEX ++#define EMAC0_DEFAULT_EMAC_MODE RGMII ++#define EMAC0_DEFAULT_EMAC_SPEED SPEED_1000M ++ ++#define EMAC1_DEFAULT_DUPLEX_MODE FULLDUPLEX ++#define EMAC1_DEFAULT_EMAC_MODE SGMII ++#define EMAC1_DEFAULT_EMAC_SPEED SPEED_1000M ++ ++/* MII-related definitios */ ++#define EMAC_MII_DATA_ST 0x40000000 /* Start of frame delimiter */ ++#define EMAC_MII_DATA_OP_RD 0x20000000 /* Perform a read operation */ ++#define EMAC_MII_DATA_OP_WR 0x10000000 /* Perform a write operation */ ++#define EMAC_MII_DATA_PA_MSK 0x0f800000 /* PHY Address field mask */ ++#define EMAC_MII_DATA_RA_MSK 0x007c0000 /* PHY Register field mask */ ++#define EMAC_MII_DATA_TA 0x00020000 /* Turnaround */ ++#define EMAC_MII_DATA_DATAMSK 0x0000ffff /* PHY data field */ ++ ++#define EMAC_MII_DATA_RA_SHIFT 18 /* MII Register address bits */ ++#define EMAC_MII_DATA_RA_MASK 0x1F /* MII Register address mask */ ++#define EMAC_MII_DATA_PA_SHIFT 23 /* MII PHY address bits */ ++#define EMAC_MII_DATA_PA_MASK 0x1F /* MII PHY address mask */ ++ ++#define EMAC_MII_DATA_RA(v) ((v & EMAC_MII_DATA_RA_MASK) << EMAC_MII_DATA_RA_SHIFT) ++#define EMAC_MII_DATA_PA(v) ((v & EMAC_MII_DATA_RA_MASK) << EMAC_MII_DATA_PA_SHIFT) ++#define EMAC_MII_DATA(v) (v & 0xffff) ++ ++#define EMAC_MII_SPEED_SHIFT 1 ++#define EMAC_HOLDTIME_SHIFT 8 ++#define EMAC_HOLDTIME_MASK 0x7 ++#define EMAC_HOLDTIME(v) ((v & EMAC_HOLDTIME_MASK) << EMAC_HOLDTIME_SHIFT) ++ ++/* The Address organisation for the MAC device. All addresses are split into ++ * two 32-bit register fields. The first one (bottom) is the lower 32-bits of ++ * the address and the other field are the high order bits - this may be 16-bits ++ * in the case of MAC addresses, or 32-bits for the hash address. ++ * In terms of memory storage, the first item (bottom) is assumed to be at a ++ * lower address location than 'top'. i.e. top should be at address location of ++ * 'bottom' + 4 bytes. ++ */ ++typedef struct { ++ u32 bottom; /* Lower 32-bits of address. */ ++ u32 top; /* Upper 32-bits of address. */ ++} MAC_ADDR; ++ ++ ++/* The following is the organisation of the address filters section of the MAC ++ * registers. The Cadence MAC contains four possible specific address match ++ * addresses, if an incoming frame corresponds to any one of these four ++ * addresses then the frame will be copied to memory. ++ * It is not necessary for all four of the address match registers to be ++ * programmed, this is application dependant. ++ */ ++typedef struct { ++ MAC_ADDR one; /* Specific address register 1. */ ++ MAC_ADDR two; /* Specific address register 2. */ ++ MAC_ADDR three; /* Specific address register 3. */ ++ MAC_ADDR four; /* Specific address register 4. */ ++} SPEC_ADDR; ++ ++typedef struct { ++ u32 mode; ++ u32 speed; ++ u32 duplex; ++} GEMAC_CFG; ++ ++/* EMAC Hash size */ ++#define EMAC_HASH_REG_BITS 64 ++ ++#define EMAC_SPEC_ADDR_MAX 4 ++ ++#endif /* _EMAC_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/gpi.h +@@ -0,0 +1,78 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _GPI_H_ ++#define _GPI_H_ ++ ++#define GPI_VERSION 0x00 ++#define GPI_CTRL 0x04 ++#define GPI_RX_CONFIG 0x08 ++#define GPI_HDR_SIZE 0x0c ++#define GPI_BUF_SIZE 0x10 ++#define GPI_LMEM_ALLOC_ADDR 0x14 ++#define GPI_LMEM_FREE_ADDR 0x18 ++#define GPI_DDR_ALLOC_ADDR 0x1c ++#define GPI_DDR_FREE_ADDR 0x20 ++#define GPI_CLASS_ADDR 0x24 ++#define GPI_DRX_FIFO 0x28 ++#define GPI_TRX_FIFO 0x2c ++#define GPI_INQ_PKTPTR 0x30 ++#define GPI_DDR_DATA_OFFSET 0x34 ++#define GPI_LMEM_DATA_OFFSET 0x38 ++#define GPI_TMLF_TX 0x4c ++#define GPI_DTX_ASEQ 0x50 ++#define GPI_FIFO_STATUS 0x54 ++#define GPI_FIFO_DEBUG 0x58 ++#define GPI_TX_PAUSE_TIME 0x5c ++#define GPI_LMEM_SEC_BUF_DATA_OFFSET 0x60 ++#define GPI_DDR_SEC_BUF_DATA_OFFSET 0x64 ++#define GPI_TOE_CHKSUM_EN 0x68 ++#define GPI_OVERRUN_DROPCNT 0x6c ++ ++typedef struct { ++ u32 lmem_rtry_cnt; ++ u32 tmlf_txthres; ++ u32 aseq_len; ++} GPI_CFG; ++ ++ ++/* GPI commons defines */ ++#define GPI_LMEM_BUF_EN 0x1 ++#define GPI_DDR_BUF_EN 0x1 ++ ++/* EGPI 1 defines */ ++#define EGPI1_LMEM_RTRY_CNT 0x40 ++#define EGPI1_TMLF_TXTHRES 0xBC ++#define EGPI1_ASEQ_LEN 0x50 ++ ++/* EGPI 2 defines */ ++#define EGPI2_LMEM_RTRY_CNT 0x40 ++#define EGPI2_TMLF_TXTHRES 0xBC ++#define EGPI2_ASEQ_LEN 0x40 ++ ++/* EGPI 3 defines */ ++#define EGPI3_LMEM_RTRY_CNT 0x40 ++#define EGPI3_TMLF_TXTHRES 0xBC ++#define EGPI3_ASEQ_LEN 0x40 ++ ++/* HGPI defines */ ++#define HGPI_LMEM_RTRY_CNT 0x40 ++#define HGPI_TMLF_TXTHRES 0xBC ++#define HGPI_ASEQ_LEN 0x40 ++ ++#endif /* _GPI_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/gpt.h +@@ -0,0 +1,29 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _CBUS_GPT_H_ ++#define _CBUS_GPT_H_ ++ ++#define CBUS_GPT_VERSION (CBUS_GPT_BASE_ADDR + 0x00) ++#define CBUS_GPT_STATUS (CBUS_GPT_BASE_ADDR + 0x04) ++#define CBUS_GPT_CONFIG (CBUS_GPT_BASE_ADDR + 0x08) ++#define CBUS_GPT_COUNTER (CBUS_GPT_BASE_ADDR + 0x0c) ++#define CBUS_GPT_PERIOD (CBUS_GPT_BASE_ADDR + 0x10) ++#define CBUS_GPT_WIDTH (CBUS_GPT_BASE_ADDR + 0x14) ++ ++#endif /* _CBUS_GPT_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/hif.h +@@ -0,0 +1,96 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _HIF_H_ ++#define _HIF_H_ ++ ++/** @file hif.h. ++ * hif - PFE hif block control and status register. Mapped on CBUS and accessible from all PE's and ARM. ++ */ ++#define HIF_VERSION (HIF_BASE_ADDR + 0x00) ++#define HIF_TX_CTRL (HIF_BASE_ADDR + 0x04) ++#define HIF_TX_CURR_BD_ADDR (HIF_BASE_ADDR + 0x08) ++#define HIF_TX_ALLOC (HIF_BASE_ADDR + 0x0c) ++#define HIF_TX_BDP_ADDR (HIF_BASE_ADDR + 0x10) ++#define HIF_TX_STATUS (HIF_BASE_ADDR + 0x14) ++#define HIF_RX_CTRL (HIF_BASE_ADDR + 0x20) ++#define HIF_RX_BDP_ADDR (HIF_BASE_ADDR + 0x24) ++#define HIF_RX_STATUS (HIF_BASE_ADDR + 0x30) ++#define HIF_INT_SRC (HIF_BASE_ADDR + 0x34) ++#define HIF_INT_ENABLE (HIF_BASE_ADDR + 0x38) ++#define HIF_POLL_CTRL (HIF_BASE_ADDR + 0x3c) ++#define HIF_RX_CURR_BD_ADDR (HIF_BASE_ADDR + 0x40) ++#define HIF_RX_ALLOC (HIF_BASE_ADDR + 0x44) ++#define HIF_TX_DMA_STATUS (HIF_BASE_ADDR + 0x48) ++#define HIF_RX_DMA_STATUS (HIF_BASE_ADDR + 0x4c) ++#define HIF_INT_COAL (HIF_BASE_ADDR + 0x50) ++ ++/*HIF_INT_SRC/ HIF_INT_ENABLE control bits */ ++#define HIF_INT (1 << 0) ++#define HIF_RXBD_INT (1 << 1) ++#define HIF_RXPKT_INT (1 << 2) ++#define HIF_TXBD_INT (1 << 3) ++#define HIF_TXPKT_INT (1 << 4) ++ ++/*HIF_TX_CTRL bits */ ++#define HIF_CTRL_DMA_EN (1<<0) ++#define HIF_CTRL_BDP_POLL_CTRL_EN (1<<1) ++#define HIF_CTRL_BDP_CH_START_WSTB (1<<2) ++ ++/*HIF_INT_ENABLE bits */ ++#define HIF_INT_EN (1 << 0) ++#define HIF_RXBD_INT_EN (1 << 1) ++#define HIF_RXPKT_INT_EN (1 << 2) ++#define HIF_TXBD_INT_EN (1 << 3) ++#define HIF_TXPKT_INT_EN (1 << 4) ++ ++/*HIF_POLL_CTRL bits*/ ++#define HIF_RX_POLL_CTRL_CYCLE 0x0400 ++#define HIF_TX_POLL_CTRL_CYCLE 0x0400 ++ ++/*HIF_INT_COAL bits*/ ++#define HIF_INT_COAL_ENABLE (1 << 31) ++ ++/*Buffer descriptor control bits */ ++#define BD_CTRL_BUFLEN_MASK 0x3fff ++#define BD_BUF_LEN(x) (x & BD_CTRL_BUFLEN_MASK) ++#define BD_CTRL_CBD_INT_EN (1 << 16) ++#define BD_CTRL_PKT_INT_EN (1 << 17) ++#define BD_CTRL_LIFM (1 << 18) ++#define BD_CTRL_LAST_BD (1 << 19) ++#define BD_CTRL_DIR (1 << 20) ++#define BD_CTRL_LMEM_CPY (1 << 21) /*Valid only for HIF_NOCPY*/ ++#define BD_CTRL_PKT_XFER (1 << 24) ++#define BD_CTRL_DESC_EN (1 << 31) ++#define BD_CTRL_PARSE_DISABLE (1 << 25) ++#define BD_CTRL_BRFETCH_DISABLE (1 << 26) ++#define BD_CTRL_RTFETCH_DISABLE (1 << 27) ++ ++/*Buffer descriptor status bits*/ ++#define BD_STATUS_CONN_ID(x) ((x) & 0xffff) ++#define BD_STATUS_DIR_PROC_ID (1 << 16) ++#define BD_STATUS_CONN_ID_EN (1 << 17)) ++#define BD_STATUS_PE2PROC_ID(x) (((x) & 7) << 18) ++#define BD_STATUS_LE_DATA (1 << 21) ++#define BD_STATUS_CHKSUM_EN (1 << 22) ++ ++/*HIF Buffer descriptor status bits */ ++#define DIR_PROC_ID (1 << 16) ++#define PROC_ID(id) ((id) << 18) ++ ++#endif /* _HIF_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/hif_nocpy.h +@@ -0,0 +1,51 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _HIF_NOCPY_H_ ++#define _HIF_NOCPY_H_ ++ ++#define HIF_NOCPY_VERSION (HIF_NOCPY_BASE_ADDR + 0x00) ++#define HIF_NOCPY_TX_CTRL (HIF_NOCPY_BASE_ADDR + 0x04) ++#define HIF_NOCPY_TX_CURR_BD_ADDR (HIF_NOCPY_BASE_ADDR + 0x08) ++#define HIF_NOCPY_TX_ALLOC (HIF_NOCPY_BASE_ADDR + 0x0c) ++#define HIF_NOCPY_TX_BDP_ADDR (HIF_NOCPY_BASE_ADDR + 0x10) ++#define HIF_NOCPY_TX_STATUS (HIF_NOCPY_BASE_ADDR + 0x14) ++#define HIF_NOCPY_RX_CTRL (HIF_NOCPY_BASE_ADDR + 0x20) ++#define HIF_NOCPY_RX_BDP_ADDR (HIF_NOCPY_BASE_ADDR + 0x24) ++#define HIF_NOCPY_RX_STATUS (HIF_NOCPY_BASE_ADDR + 0x30) ++#define HIF_NOCPY_INT_SRC (HIF_NOCPY_BASE_ADDR + 0x34) ++#define HIF_NOCPY_INT_ENABLE (HIF_NOCPY_BASE_ADDR + 0x38) ++#define HIF_NOCPY_POLL_CTRL (HIF_NOCPY_BASE_ADDR + 0x3c) ++#define HIF_NOCPY_RX_CURR_BD_ADDR (HIF_NOCPY_BASE_ADDR + 0x40) ++#define HIF_NOCPY_RX_ALLOC (HIF_NOCPY_BASE_ADDR + 0x44) ++#define HIF_NOCPY_TX_DMA_STATUS (HIF_NOCPY_BASE_ADDR + 0x48) ++#define HIF_NOCPY_RX_DMA_STATUS (HIF_NOCPY_BASE_ADDR + 0x4c) ++#define HIF_NOCPY_RX_INQ0_PKTPTR (HIF_NOCPY_BASE_ADDR + 0x50) ++#define HIF_NOCPY_RX_INQ1_PKTPTR (HIF_NOCPY_BASE_ADDR + 0x54) ++#define HIF_NOCPY_TX_PORT_NO (HIF_NOCPY_BASE_ADDR + 0x60) ++#define HIF_NOCPY_LMEM_ALLOC_ADDR (HIF_NOCPY_BASE_ADDR + 0x64) ++#define HIF_NOCPY_CLASS_ADDR (HIF_NOCPY_BASE_ADDR + 0x68) ++#define HIF_NOCPY_TMU_PORT0_ADDR (HIF_NOCPY_BASE_ADDR + 0x70) ++#define HIF_NOCPY_TMU_PORT1_ADDR (HIF_NOCPY_BASE_ADDR + 0x74) ++#define HIF_NOCPY_TMU_PORT2_ADDR (HIF_NOCPY_BASE_ADDR + 0x7c) ++#define HIF_NOCPY_TMU_PORT3_ADDR (HIF_NOCPY_BASE_ADDR + 0x80) ++#define HIF_NOCPY_TMU_PORT4_ADDR (HIF_NOCPY_BASE_ADDR + 0x84) ++#define HIF_NOCPY_INT_COAL (HIF_NOCPY_BASE_ADDR + 0x90) ++ ++ ++#endif /* _HIF_NOCPY_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/tmu_csr.h +@@ -0,0 +1,128 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _TMU_CSR_H_ ++#define _TMU_CSR_H_ ++ ++#define TMU_VERSION (TMU_CSR_BASE_ADDR + 0x000) ++#define TMU_INQ_WATERMARK (TMU_CSR_BASE_ADDR + 0x004) ++#define TMU_PHY_INQ_PKTPTR (TMU_CSR_BASE_ADDR + 0x008) ++#define TMU_PHY_INQ_PKTINFO (TMU_CSR_BASE_ADDR + 0x00c) ++#define TMU_PHY_INQ_FIFO_CNT (TMU_CSR_BASE_ADDR + 0x010) ++#define TMU_SYS_GENERIC_CONTROL (TMU_CSR_BASE_ADDR + 0x014) ++#define TMU_SYS_GENERIC_STATUS (TMU_CSR_BASE_ADDR + 0x018) ++#define TMU_SYS_GEN_CON0 (TMU_CSR_BASE_ADDR + 0x01c) ++#define TMU_SYS_GEN_CON1 (TMU_CSR_BASE_ADDR + 0x020) ++#define TMU_SYS_GEN_CON2 (TMU_CSR_BASE_ADDR + 0x024) ++#define TMU_SYS_GEN_CON3 (TMU_CSR_BASE_ADDR + 0x028) ++#define TMU_SYS_GEN_CON4 (TMU_CSR_BASE_ADDR + 0x02c) ++#define TMU_TEQ_DISABLE_DROPCHK (TMU_CSR_BASE_ADDR + 0x030) ++#define TMU_TEQ_CTRL (TMU_CSR_BASE_ADDR + 0x034) ++#define TMU_TEQ_QCFG (TMU_CSR_BASE_ADDR + 0x038) ++#define TMU_TEQ_DROP_STAT (TMU_CSR_BASE_ADDR + 0x03c) ++#define TMU_TEQ_QAVG (TMU_CSR_BASE_ADDR + 0x040) ++#define TMU_TEQ_WREG_PROB (TMU_CSR_BASE_ADDR + 0x044) ++#define TMU_TEQ_TRANS_STAT (TMU_CSR_BASE_ADDR + 0x048) ++#define TMU_TEQ_HW_PROB_CFG0 (TMU_CSR_BASE_ADDR + 0x04c) ++#define TMU_TEQ_HW_PROB_CFG1 (TMU_CSR_BASE_ADDR + 0x050) ++#define TMU_TEQ_HW_PROB_CFG2 (TMU_CSR_BASE_ADDR + 0x054) ++#define TMU_TEQ_HW_PROB_CFG3 (TMU_CSR_BASE_ADDR + 0x058) ++#define TMU_TEQ_HW_PROB_CFG4 (TMU_CSR_BASE_ADDR + 0x05c) ++#define TMU_TEQ_HW_PROB_CFG5 (TMU_CSR_BASE_ADDR + 0x060) ++#define TMU_TEQ_HW_PROB_CFG6 (TMU_CSR_BASE_ADDR + 0x064) ++#define TMU_TEQ_HW_PROB_CFG7 (TMU_CSR_BASE_ADDR + 0x068) ++#define TMU_TEQ_HW_PROB_CFG8 (TMU_CSR_BASE_ADDR + 0x06c) ++#define TMU_TEQ_HW_PROB_CFG9 (TMU_CSR_BASE_ADDR + 0x070) ++#define TMU_TEQ_HW_PROB_CFG10 (TMU_CSR_BASE_ADDR + 0x074) ++#define TMU_TEQ_HW_PROB_CFG11 (TMU_CSR_BASE_ADDR + 0x078) ++#define TMU_TEQ_HW_PROB_CFG12 (TMU_CSR_BASE_ADDR + 0x07c) ++#define TMU_TEQ_HW_PROB_CFG13 (TMU_CSR_BASE_ADDR + 0x080) ++#define TMU_TEQ_HW_PROB_CFG14 (TMU_CSR_BASE_ADDR + 0x084) ++#define TMU_TEQ_HW_PROB_CFG15 (TMU_CSR_BASE_ADDR + 0x088) ++#define TMU_TEQ_HW_PROB_CFG16 (TMU_CSR_BASE_ADDR + 0x08c) ++#define TMU_TEQ_HW_PROB_CFG17 (TMU_CSR_BASE_ADDR + 0x090) ++#define TMU_TEQ_HW_PROB_CFG18 (TMU_CSR_BASE_ADDR + 0x094) ++#define TMU_TEQ_HW_PROB_CFG19 (TMU_CSR_BASE_ADDR + 0x098) ++#define TMU_TEQ_HW_PROB_CFG20 (TMU_CSR_BASE_ADDR + 0x09c) ++#define TMU_TEQ_HW_PROB_CFG21 (TMU_CSR_BASE_ADDR + 0x0a0) ++#define TMU_TEQ_HW_PROB_CFG22 (TMU_CSR_BASE_ADDR + 0x0a4) ++#define TMU_TEQ_HW_PROB_CFG23 (TMU_CSR_BASE_ADDR + 0x0a8) ++#define TMU_TEQ_HW_PROB_CFG24 (TMU_CSR_BASE_ADDR + 0x0ac) ++#define TMU_TEQ_HW_PROB_CFG25 (TMU_CSR_BASE_ADDR + 0x0b0) ++#define TMU_TDQ_IIFG_CFG (TMU_CSR_BASE_ADDR + 0x0b4) ++#define TMU_TDQ0_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x0b8) /**< [9:0] Scheduler Enable for each of the scheduler in the TDQ. This is a global Enable for all schedulers in PHY0 */ ++#define TMU_LLM_CTRL (TMU_CSR_BASE_ADDR + 0x0bc) ++#define TMU_LLM_BASE_ADDR (TMU_CSR_BASE_ADDR + 0x0c0) ++#define TMU_LLM_QUE_LEN (TMU_CSR_BASE_ADDR + 0x0c4) ++#define TMU_LLM_QUE_HEADPTR (TMU_CSR_BASE_ADDR + 0x0c8) ++#define TMU_LLM_QUE_TAILPTR (TMU_CSR_BASE_ADDR + 0x0cc) ++#define TMU_LLM_QUE_DROPCNT (TMU_CSR_BASE_ADDR + 0x0d0) ++#define TMU_INT_EN (TMU_CSR_BASE_ADDR + 0x0d4) ++#define TMU_INT_SRC (TMU_CSR_BASE_ADDR + 0x0d8) ++#define TMU_INQ_STAT (TMU_CSR_BASE_ADDR + 0x0dc) ++#define TMU_CTRL (TMU_CSR_BASE_ADDR + 0x0e0) ++ ++#define TMU_MEM_ACCESS_ADDR (TMU_CSR_BASE_ADDR + 0x0e4) /**< [31] Mem Access Command. 0 = Internal Memory Read, 1 = Internal memory Write [27:24] Byte Enables of the Internal memory access [23:0] Address of the internal memory. This address is used to access both the PM and DM of all the PE's */ ++#define TMU_MEM_ACCESS_WDATA (TMU_CSR_BASE_ADDR + 0x0e8) /**< Internal Memory Access Write Data */ ++#define TMU_MEM_ACCESS_RDATA (TMU_CSR_BASE_ADDR + 0x0ec) /**< Internal Memory Access Read Data. The commands are blocked at the mem_access only */ ++ ++#define TMU_PHY0_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0f0) /**< [31:0] PHY0 in queue address (must be initialized with one of the xxx_INQ_PKTPTR cbus addresses) */ ++#define TMU_PHY1_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0f4) /**< [31:0] PHY1 in queue address (must be initialized with one of the xxx_INQ_PKTPTR cbus addresses) */ ++#define TMU_PHY2_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0f8) /**< [31:0] PHY2 in queue address (must be initialized with one of the xxx_INQ_PKTPTR cbus addresses) */ ++#define TMU_PHY3_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x0fc) /**< [31:0] PHY3 in queue address (must be initialized with one of the xxx_INQ_PKTPTR cbus addresses) */ ++#define TMU_BMU_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x100) ++#define TMU_TX_CTRL (TMU_CSR_BASE_ADDR + 0x104) ++ ++#define TMU_BUS_ACCESS_WDATA (TMU_CSR_BASE_ADDR + 0x108) ++#define TMU_BUS_ACCESS (TMU_CSR_BASE_ADDR + 0x10c) ++#define TMU_BUS_ACCESS_RDATA (TMU_CSR_BASE_ADDR + 0x110) ++ ++#define TMU_PE_SYS_CLK_RATIO (TMU_CSR_BASE_ADDR + 0x114) ++#define TMU_PE_STATUS (TMU_CSR_BASE_ADDR + 0x118) ++#define TMU_TEQ_MAX_THRESHOLD (TMU_CSR_BASE_ADDR + 0x11c) ++#define TMU_PHY4_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x134) /**< [31:0] PHY4 in queue address (must be initialized with one of the xxx_INQ_PKTPTR cbus addresses) */ ++#define TMU_TDQ1_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x138) /**< [9:0] Scheduler Enable for each of the scheduler in the TDQ. This is a global Enable for all schedulers in PHY1 */ ++#define TMU_TDQ2_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x13c) /**< [9:0] Scheduler Enable for each of the scheduler in the TDQ. This is a global Enable for all schedulers in PHY2 */ ++#define TMU_TDQ3_SCH_CTRL (TMU_CSR_BASE_ADDR + 0x140) /**< [9:0] Scheduler Enable for each of the scheduler in the TDQ. This is a global Enable for all schedulers in PHY3 */ ++#define TMU_BMU_BUF_SIZE (TMU_CSR_BASE_ADDR + 0x144) ++#define TMU_PHY5_INQ_ADDR (TMU_CSR_BASE_ADDR + 0x148) /**< [31:0] PHY5 in queue address (must be initialized with one of the xxx_INQ_PKTPTR cbus addresses) */ ++ ++#define SW_RESET (1 << 0) /**< Global software reset */ ++#define INQ_RESET (1 << 2) ++#define TEQ_RESET (1 << 3) ++#define TDQ_RESET (1 << 4) ++#define PE_RESET (1 << 5) ++#define MEM_INIT (1 << 6) ++#define MEM_INIT_DONE (1 << 7) ++#define LLM_INIT (1 << 8) ++#define LLM_INIT_DONE (1 << 9) ++#define ECC_MEM_INIT_DONE (1<<10) ++ ++typedef struct { ++ u32 pe_sys_clk_ratio; ++ unsigned long llm_base_addr; ++ u32 llm_queue_len; ++} TMU_CFG; ++ ++/* Not HW related for pfe_ctrl / pfe common defines */ ++#define DEFAULT_MAX_QDEPTH 80 ++#define DEFAULT_Q0_QDEPTH 511 //We keep one large queue for host tx qos ++#define DEFAULT_TMU3_QDEPTH 127 ++ ++ ++#endif /* _TMU_CSR_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/cbus/util_csr.h +@@ -0,0 +1,61 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _UTIL_CSR_H_ ++#define _UTIL_CSR_H_ ++ ++#define UTIL_VERSION (UTIL_CSR_BASE_ADDR + 0x000) ++#define UTIL_TX_CTRL (UTIL_CSR_BASE_ADDR + 0x004) ++#define UTIL_INQ_PKTPTR (UTIL_CSR_BASE_ADDR + 0x010) ++ ++#define UTIL_HDR_SIZE (UTIL_CSR_BASE_ADDR + 0x014) ++ ++#define UTIL_PE0_QB_DM_ADDR0 (UTIL_CSR_BASE_ADDR + 0x020) ++#define UTIL_PE0_QB_DM_ADDR1 (UTIL_CSR_BASE_ADDR + 0x024) ++#define UTIL_PE0_RO_DM_ADDR0 (UTIL_CSR_BASE_ADDR + 0x060) ++#define UTIL_PE0_RO_DM_ADDR1 (UTIL_CSR_BASE_ADDR + 0x064) ++ ++#define UTIL_MEM_ACCESS_ADDR (UTIL_CSR_BASE_ADDR + 0x100) ++#define UTIL_MEM_ACCESS_WDATA (UTIL_CSR_BASE_ADDR + 0x104) ++#define UTIL_MEM_ACCESS_RDATA (UTIL_CSR_BASE_ADDR + 0x108) ++ ++#define UTIL_TM_INQ_ADDR (UTIL_CSR_BASE_ADDR + 0x114) ++#define UTIL_PE_STATUS (UTIL_CSR_BASE_ADDR + 0x118) ++ ++#define UTIL_PE_SYS_CLK_RATIO (UTIL_CSR_BASE_ADDR + 0x200) ++#define UTIL_AFULL_THRES (UTIL_CSR_BASE_ADDR + 0x204) ++#define UTIL_GAP_BETWEEN_READS (UTIL_CSR_BASE_ADDR + 0x208) ++#define UTIL_MAX_BUF_CNT (UTIL_CSR_BASE_ADDR + 0x20c) ++#define UTIL_TSQ_FIFO_THRES (UTIL_CSR_BASE_ADDR + 0x210) ++#define UTIL_TSQ_MAX_CNT (UTIL_CSR_BASE_ADDR + 0x214) ++#define UTIL_IRAM_DATA_0 (UTIL_CSR_BASE_ADDR + 0x218) ++#define UTIL_IRAM_DATA_1 (UTIL_CSR_BASE_ADDR + 0x21c) ++#define UTIL_IRAM_DATA_2 (UTIL_CSR_BASE_ADDR + 0x220) ++#define UTIL_IRAM_DATA_3 (UTIL_CSR_BASE_ADDR + 0x224) ++ ++#define UTIL_BUS_ACCESS_ADDR (UTIL_CSR_BASE_ADDR + 0x228) ++#define UTIL_BUS_ACCESS_WDATA (UTIL_CSR_BASE_ADDR + 0x22c) ++#define UTIL_BUS_ACCESS_RDATA (UTIL_CSR_BASE_ADDR + 0x230) ++ ++#define UTIL_INQ_AFULL_THRES (UTIL_CSR_BASE_ADDR + 0x234) ++ ++typedef struct { ++ u32 pe_sys_clk_ratio; ++} UTIL_CFG; ++ ++#endif /* _UTIL_CSR_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/class.h +@@ -0,0 +1,133 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _CLASS_H_ ++#define _CLASS_H_ ++ ++#include "pe.h" ++ ++#define CLASS_DMEM_BASE_ADDR 0x00000000 ++#define CLASS_DMEM_SIZE 0x2000 ++#define CLASS_DMEM_END (CLASS_DMEM_BASE_ADDR + CLASS_DMEM_SIZE) ++#define CLASS_PMEM_BASE_ADDR 0x00010000 ++ ++#define CBUS_BASE_ADDR 0xc0000000 ++#define CLASS_APB_BASE_ADDR 0xc1000000 ++#define CLASS_AHB1_BASE_ADDR 0xc2000000 ++#define CLASS_AHB2_BASE_ADDR 0xc3000000 ++ ++#include "cbus.h" ++ ++#define GPT_BASE_ADDR (CLASS_APB_BASE_ADDR + 0x00000) ++#define UART_BASE_ADDR (CLASS_APB_BASE_ADDR + 0x10000) ++#define PERG_BASE_ADDR (CLASS_APB_BASE_ADDR + 0x20000) ++#define EFET_BASE_ADDR (CLASS_APB_BASE_ADDR + 0x40000) ++ ++#define MAC_HASH_BASE_ADDR (CLASS_AHB1_BASE_ADDR + 0x30000) ++#define VLAN_HASH_BASE_ADDR (CLASS_AHB1_BASE_ADDR + 0x50000) ++ ++#define PE_LMEM_BASE_ADDR (CLASS_AHB2_BASE_ADDR + 0x10000) ++#define PE_LMEM_SIZE 0x8000 ++#define PE_LMEM_END (PE_LMEM_BASE_ADDR + PE_LMEM_SIZE) ++#define CCU_BASE_ADDR (CLASS_AHB2_BASE_ADDR + 0x20000) ++ ++#define IS_DMEM(addr, len) (((unsigned long)(addr) >= CLASS_DMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= CLASS_DMEM_END)) ++#define IS_PE_LMEM(addr, len) (((unsigned long)(addr) >= PE_LMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= PE_LMEM_END)) ++ ++ ++#include "gpt.h" ++#include "uart.h" ++#include "class/perg.h" ++#include "class/efet.h" ++#include "class/mac_hash.h" ++#include "class/vlan_hash.h" ++#include "class/ccu.h" ++ ++ ++#define CLASS_MAX_PBUFFERS 4 ++ ++#define PBUF_HWPARSE_OFFSET 0x10 /* Fixed by hardware */ ++ ++#define PAYLOAD_DMEM_MAX_SIZE (CLASS_PBUF_SIZE - CLASS_PBUF_HEADER_OFFSET - sizeof(class_rx_hdr_t)) ++ ++ ++#define MIN_PKT_SIZE 56 ++ ++#define PARSE_ETH_TYPE (1 << 0) ++#define PARSE_VLAN_TYPE (1 << 1) ++#define PARSE_PPPOE_TYPE (1 << 2) ++#define PARSE_ARP_TYPE (1 << 3) ++#define PARSE_MCAST_TYPE (1 << 4) ++#define PARSE_IP_TYPE (1 << 5) ++#define PARSE_IPV6_TYPE (1 << 6) ++#define PARSE_IPV4_TYPE (1 << 7) ++ ++#define PARSE_IPX_TYPE (1 << 9) ++ ++#define PARSE_UDP_FLOW (1 << 11) ++#define PARSE_TCP_FLOW (1 << 12) ++#define PARSE_ICMP_FLOW (1 << 13) ++#define PARSE_IGMP_FLOW (1 << 14) ++#define PARSE_FRAG_FLOW (1 << 15) ++ ++#define PARSE_HIF_PKT (1 << 23) ++#define PARSE_ARC_HIT (1 << 24) ++#define PARSE_PKT_OVERFLOW (1 << 25) ++ ++#define PARSE_PROTO_MISMATCH (1 << 28) ++#define PARSE_L3_MISMATCH (1 << 29) ++#define PARSE_L2_MISMATCH (1 << 30) ++#define PARSE_INCOMPLETE (1 << 31) ++ ++ ++typedef struct _hwparse_t { ++ u16 sid; ++ u16 connid; ++ u8 toevec; ++ u8 pLayer2Hdr; ++ u8 pLayer3Hdr; ++ u8 pLayer4Hdr; ++ u16 vlanid; ++ u16 ifParseFlags; ++ u32 parseFlags; ++ u16 srcport; ++ u16 dstport; ++ u32 proto:8; ++ u32 port:4; ++ u32 hash:20; ++ u64 rte_res_valid:1; ++ u64 vlan_res_valid:1; ++ u64 dst_res_valid:1; ++ u64 src_res_valid:1; ++ u64 vlan_lookup:20; ++ u64 dst_lookup:20; ++ u64 src_lookup:20; ++} hwparse_t; ++ ++ ++typedef struct { ++ u8 num_cpy; /* no of copies to send out from RO block, for each there must be a corresponding tx pre-header */ ++ u8 dma_len; /* len to be DMAed to DDR mem, including all tx pre-headers */ ++ u16 src_addr; /* class dmem source address, pointing to first tx pre-header */ ++ u32 dst_addr; /* DDR memory destination address of first tx pre-header, must be so packet data is continuous in DDR */ ++ u32 res1; /* reserved for software usage - queue number? */ ++ u16 res2; /* reserved for software usage */ ++ u16 tsv; /* time stamp val */ ++} class_tx_desc_t; ++ ++#endif /* _CLASS_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/class/ccu.h +@@ -0,0 +1,28 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _CCU_H_ ++#define _CCU_H_ ++ ++#define CCU_ADDR (CCU_BASE_ADDR + 0x00) ++#define CCU_CNT (CCU_BASE_ADDR + 0x04) ++#define CCU_STATUS (CCU_BASE_ADDR + 0x08) ++#define CCU_VAL (CCU_BASE_ADDR + 0x0c) ++ ++#endif /* _CCU_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/class/efet.h +@@ -0,0 +1,44 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _CLASS_EFET_H_ ++#define _CLASS_EFET_H_ ++ ++//#define CLASS_EFET_ASYNC 1 ++ ++#define CLASS_EFET_ENTRY_ADDR (EFET_BASE_ADDR + 0x00) ++#define CLASS_EFET_ENTRY_SIZE (EFET_BASE_ADDR + 0x04) ++#define CLASS_EFET_ENTRY_DMEM_ADDR (EFET_BASE_ADDR + 0x08) ++#define CLASS_EFET_ENTRY_STATUS (EFET_BASE_ADDR + 0x0c) ++#define CLASS_EFET_ENTRY_ENDIAN (EFET_BASE_ADDR + 0x10) ++ ++#define CBUS2DMEM 0 ++#define DMEM2CBUS 1 ++ ++#define EFET2BUS_LE (1 << 0) ++#define PE2BUS_LE (1 << 1) ++ ++#ifdef CLASS_EFET_ASYNC ++void class_efet_async(u32 cbus_addr, u32 dmem_addr, u32 len, u32 dir); ++#endif ++ ++void class_efet_sync(u32 cbus_addr, u32 dmem_addr, u32 len, u32 dir); ++ ++ ++#endif /* _CLASS_EFET_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/class/mac_hash.h +@@ -0,0 +1,55 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _MAC_HASH_H_ ++#define _MAC_HASH_H_ ++ ++#define MAC_HASH_REQ1_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x000) ++#define MAC_HASH_REQ2_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x020) ++#define MAC_HASH_REQ3_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x040) ++#define MAC_HASH_REQ4_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x060) ++#define MAC_HASH_REQ5_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x080) ++#define MAC_HASH_REQ6_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x0a0) ++#define MAC_HASH_REQ7_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x0c0) ++#define MAC_HASH_REQ8_BASE_ADDR (MAC_HASH_BASE_ADDR + 0x0e0) ++ ++#define MAC_HASH_REQ_CMD(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x000) ++#define MAC_HASH_REQ_MAC1_ADDR(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x004) ++#define MAC_HASH_REQ_MAC2_ADDR(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x008) ++#define MAC_HASH_REQ_MASK1_ADDR(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x00c) ++#define MAC_HASH_REQ_MASK2_ADDR(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x010) ++#define MAC_HASH_REQ_ENTRY(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x014) ++#define MAC_HASH_REQ_STATUS(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x018) ++#define MAC_HASH_REQ_ENTRY_MAYCH(i) (MAC_HASH_REQ##i##_BASE_ADDR + 0x01c) ++ ++ ++#define MAC_HASH_FREELIST_PTR_HEAD (MAC_HASH_BASE_ADDR + 0x100) ++#define MAC_HASH_FREELIST_PTR_TAIL (MAC_HASH_BASE_ADDR + 0x104) ++#define MAC_HASH_FREELIST_ENTRIES_ADDR (MAC_HASH_BASE_ADDR + 0x108) ++ ++ ++#define HASH_CMD_INIT 1 ++#define HASH_CMD_ADD 2 ++#define HASH_CMD_DELETE 3 ++#define HASH_CMD_UPDATE 4 ++#define HASH_CMD_SEARCH 5 ++#define HASH_CMD_MEM_READ 6 ++#define HASH_CMD_MEM_WRITE 7 ++ ++#endif /* _MAC_HASH_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/class/perg.h +@@ -0,0 +1,39 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _PERG_H_ ++#define _PERG_H_ ++ ++#define PERG_QB_BUF_STATUS (PERG_BASE_ADDR + 0x00) ++#define PERG_RO_BUF_STATUS (PERG_BASE_ADDR + 0x04) ++#define PERG_CLR_QB_BUF_STATUS (PERG_BASE_ADDR + 0x08) ++#define PERG_SET_RO_BUF_STATUS (PERG_BASE_ADDR + 0x0c) ++#define PERG_CLR_RO_ERR_PKT (PERG_BASE_ADDR + 0x10) ++#define PERG_CLR_BMU2_ERR_PKT (PERG_BASE_ADDR + 0x14) ++ ++#define PERG_ID (PERG_BASE_ADDR + 0x18) ++#define PERG_TIMER1 (PERG_BASE_ADDR + 0x1c) ++//FIXME #define PERG_TIMER2 (PERG_BASE_ADDR + 0x20) ++#define PERG_BMU1_CURRDEPTH (PERG_BASE_ADDR + 0x20) ++#define PERG_BMU2_CURRDEPTH (PERG_BASE_ADDR + 0x24) ++#define PERG_HOST_GP (PERG_BASE_ADDR + 0x2c) ++#define PERG_PE_GP (PERG_BASE_ADDR + 0x30) ++#define PERG_INT_ENABLE (PERG_BASE_ADDR + 0x34) ++#define PERG_INT_SRC (PERG_BASE_ADDR + 0x38) ++ ++#endif /* _PERG_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/class/vlan_hash.h +@@ -0,0 +1,46 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _VLAN_HASH_H_ ++#define _VLAN_HASH_H_ ++ ++#define VLAN_HASH_REQ1_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x000) ++#define VLAN_HASH_REQ2_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x020) ++#define VLAN_HASH_REQ3_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x040) ++#define VLAN_HASH_REQ4_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x060) ++#define VLAN_HASH_REQ5_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x080) ++#define VLAN_HASH_REQ6_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x0a0) ++#define VLAN_HASH_REQ7_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x0c0) ++#define VLAN_HASH_REQ8_BASE_ADDR (VLAN_HASH_BASE_ADDR + 0x0e0) ++ ++#define VLAN_HASH_REQ_CMD(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x000) ++#define VLAN_HASH_REQ_MAC1_ADDR(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x004) ++#define VLAN_HASH_REQ_MAC2_ADDR(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x008) ++#define VLAN_HASH_REQ_MASK1_ADDR(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x00c) ++#define VLAN_HASH_REQ_MASK2_ADDR(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x010) ++#define VLAN_HASH_REQ_ENTRY(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x014) ++#define VLAN_HASH_REQ_STATUS(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x018) ++#define VLAN_HASH_REQ_ENTRY_MAYCH(i) (VLAN_HASH_REQ##i##_BASE_ADDR + 0x01c) ++ ++ ++#define VLAN_HASH_FREELIST_PTR_HEAD (VLAN_HASH_BASE_ADDR + 0x100) ++#define VLAN_HASH_FREELIST_PTR_TAIL (VLAN_HASH_BASE_ADDR + 0x104) ++#define VLAN_HASH_FREELIST_ENTRIES_ADDR (VLAN_HASH_BASE_ADDR + 0x108) ++ ++#endif /* _VLAN_HASH_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/gpt.h +@@ -0,0 +1,44 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _GPT_H_ ++#define _GPT_H_ ++ ++#define GPT_VERSION (GPT_BASE_ADDR + 0x00) ++#define GPT_STATUS (GPT_BASE_ADDR + 0x04) ++#define GPT_CONFIG (GPT_BASE_ADDR + 0x08) ++#define GPT_COUNTER (GPT_BASE_ADDR + 0x0c) ++#define GPT_PERIOD (GPT_BASE_ADDR + 0x10) ++#define GPT_WIDTH (GPT_BASE_ADDR + 0x14) ++ ++/*** These bits are defined for GPT_STATUS register */ ++#define GPT_STAT_IRQ (1<<0) ++#define GPT_STAT_OVERFLOW_ERR (1<<4) ++#define GPT_STAT_TMR_ENABLE (1<<8) ++#define GPT_STAT_TMR_DISABLE (1<<9) ++ ++/*** These bits are defined for GPT_CONFIG register */ ++#define GPT_CONFIG_PWM_MODE 0x1 ++#define GPT_CONFIG_WCAP_MODE 0x2 ++#define GPT_CONFIG_CAP_PULSE_OUT (1<<2) ++#define GPT_CONFIG_PERIOD_CNT (1<<3) ++#define GPT_CONFIG_INTR_ENABLE (1<<4) ++#define GPT_CONFIG_AUX_SEL (1<<5) ++ ++ ++#endif /* _GPT_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/pe.h +@@ -0,0 +1,626 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _PE_H_ ++#define _PE_H_ ++ ++#include "hal.h" ++ ++#if defined(COMCERTO_2000_CLASS) ++#include "pfe/class.h" ++#elif defined(COMCERTO_2000_TMU) ++#include "pfe/tmu.h" ++#elif defined(COMCERTO_2000_UTIL) ++#include "pfe/util.h" ++#endif ++ ++enum { ++ CLASS0_ID = 0, ++ CLASS1_ID, ++ CLASS2_ID, ++ CLASS3_ID, ++ CLASS4_ID, ++ CLASS5_ID, ++ TMU0_ID, ++ TMU1_ID, ++ TMU2_ID, ++ TMU3_ID, ++ UTIL_ID, ++ MAX_PE ++}; ++#define PE_ID_ANY MAX_PE ++ ++/* Hardware definition of physical ports */ ++/* CLASS rx header phy number */ ++enum CLASS_RX_PHY { ++ RX_PHY_0 = 0x0, ++ RX_PHY_1, ++ RX_PHY_2, ++ RX_PHY_HIF, ++ RX_PHY_HIF_NOCPY, ++ RX_PHY_CLASS = 1 << 14, /**< Control bit (in PHYNO field) used to inform CLASS PE that packet comes from Class. */ ++ RX_PHY_UTIL = 1 << 15 /**< Control bit (in PHYNO field) used to inform CLASS PE that packet comes from UtilPE. */ ++}; ++ ++#define RX_PHY_SW_INPUT_PORT_OFFSET 11 /**< Offset in PHYNO field where the original input port will be stored for packets coming directly from software (UtilPE or Class). */ ++ ++ ++/* CLASS/TMU tx header phy number */ ++enum TMU_TX_PHY { ++ TX_PHY_TMU0 = 0x0, ++ TX_PHY_TMU1, ++ TX_PHY_TMU2, ++ TX_PHY_TMU3 ++}; ++ ++ ++// NOTE: Any changes to the following drop counter definitions must also ++// be reflected in the pfe/pfe.h file and in pfe_ctrl/pfe_sysfs.c. ++ ++#if defined(COMCERTO_2000_CLASS) ++ ++#define CLASS_DROP_ICC 0 ++#define CLASS_DROP_HOST_PKT_ERROR 1 ++#define CLASS_DROP_RX_ERROR 2 ++#define CLASS_DROP_IPSEC_OUT 3 ++#define CLASS_DROP_IPSEC_IN 4 ++#define CLASS_DROP_EXPT_IPSEC 5 ++#define CLASS_DROP_REASSEMBLY 6 ++#define CLASS_DROP_FRAGMENTER 7 ++#define CLASS_DROP_NATT 8 ++#define CLASS_DROP_SOCKET 9 ++#define CLASS_DROP_MULTICAST 10 ++#define CLASS_DROP_NATPT 11 ++#define CLASS_DROP_TX_DISABLE 12 ++ ++#define CLASS_NUM_DROP_COUNTERS 13 ++ ++extern U32 drop_counter[CLASS_NUM_DROP_COUNTERS]; ++#define DROP_PACKET(pmtd, counter) free_packet(pmtd, CLASS_DROP_##counter) ++#define DROP_BUFFER(addr, counter) free_buffer(addr, CLASS_DROP_##counter) ++ ++#elif defined(COMCERTO_2000_UTIL) ++ ++#define UTIL_DROP_IPSEC_OUT 0 ++#define UTIL_DROP_IPSEC_IN 1 ++#define UTIL_DROP_IPSEC_RATE_LIMIT 2 ++#define UTIL_DROP_FRAGMENTER 3 ++#define UTIL_DROP_SOCKET 4 ++#define UTIL_DROP_TX_DISABLE 5 ++#define UTIL_DROP_RX_ERROR 6 ++#define UTIL_DROP_NO_MTD 7 ++ ++#define UTIL_NUM_DROP_COUNTERS 8 ++ ++extern U32 drop_counter[UTIL_NUM_DROP_COUNTERS]; ++#define DROP_PACKET(pmtd, counter) free_packet(pmtd, UTIL_DROP_##counter) ++#define DROP_BUFFER(addr, counter) free_buffer(addr, UTIL_DROP_##counter) ++ ++#endif ++ ++ ++ ++#define DDR_BASE_ADDR 0x00020000 ++#define DDR_END 0x86000000 /* This includes ACP and IRAM areas */ ++#define IRAM_BASE_ADDR 0x83000000 ++ ++#define IS_DDR(addr, len) (((unsigned long)(addr) >= DDR_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= DDR_END)) ++/* action bits of act_phyno is defined as follows */ ++ ++#define ACT_SRC_MAC_REPLACE (1 << (4 + 0)) ++#define ACT_VLAN_ADD (1 << (4 + 1)) ++#define ACT_TCPCHKSUM_REPLACE (1 << (4 + 2)) ++#define ACT_VLAN_REPLACE (1 << (4 + 3)) ++#define ACT_DONT_FREE_BUFFER (1 << (4 + 5)) ++#define ACT_IPCHKSUM_REPLACE (1 << (4 + 6)) ++ ++typedef struct { ++ u8 start_data_off; /* packet data start offset, relative to start of this tx pre-header */ ++ u8 start_buf_off; /* this tx pre-header start offset, relative to start of DDR buffer */ ++ u16 pkt_length; /* total packet length */ ++ u8 act_phyno; /* action / phy number */ ++ u8 queueno; /* queueno */ ++ u16 unused; ++} class_tx_hdr_t; ++ ++typedef struct { ++ u8 start_data_off; /* packet data start offset, relative to start of this tx pre-header */ ++ u8 start_buf_off; /* this tx pre-header start offset, relative to start of DDR buffer */ ++ u16 pkt_length; /* total packet length */ ++ u8 act_phyno; /* action / phy number */ ++ u8 queueno; /* queueno */ ++ u16 src_mac_msb; /* indicates src_mac 47:32 */ ++ u32 src_mac_lsb; /* indicates src_mac 31:0 */ ++ u32 vlanid; /* vlanid */ ++} class_tx_hdr_mc_t; ++ ++typedef struct { ++ u32 next_ptr; /* ptr to the start of the first DDR buffer */ ++ u16 length; /* total packet length */ ++ u16 phyno; /* input physical port number */ ++ u32 status; /* gemac status bits bits[32:63]*/ ++ u32 status2; /* gemac status bits bits[0:31] */ ++} class_rx_hdr_t; ++/* class_rx_hdr status bits (status0 bits in hardware blocks) ++ * from hif_top/dma_dxr_dtx.v ++ * STATUS[9:0] is the encoding of bits in the LMEM buffer as seen by the QB block, ++ * NOT the encoding of bits as seen by the Class PEs in the DMEM rx header */ ++#define STATUS_PARSE_DISABLE (1 << 0) ++#define STATUS_BRFETCH_DISABLE (1 << 1) ++#define STATUS_RTFETCH_DISABLE (1 << 2) ++#define STATUS_DIR_PROC_ID (1 << 3) ++#define STATUS_CONN_ID_EN (1 << 4)) ++#define STATUS_PE2PROC_ID(x) (((x) & 7) << 5) ++#define STATUS_LE_DATA (1 << 8) ++#define STATUS_CHKSUM_EN (1 << 9) ++ ++/* from gpi/gpi_rmlf.v */ ++#define STATUS_CUMULATIVE_ERR (1 << 16) ++#define STATUS_LENGTH_ERR (1 << 17) ++#define STATUS_CRC_ERR (1 << 18) ++#define STATUS_TOO_SHORT_ERR (1 << 19) ++#define STATUS_TOO_LONG_ERR (1 << 20) ++#define STATUS_CODE_ERR (1 << 21) ++#define STATUS_MC_HASH_MATCH (1 << 22) ++#define STATUS_CUMULATIVE_ARC_HIT (1 << 23) ++#define STATUS_UNICAST_HASH_MATCH (1 << 24) ++#define STATUS_IP_CHECKSUM_CORRECT (1 << 25) ++#define STATUS_TCP_CHECKSUM_CORRECT (1 << 26) ++#define STATUS_UDP_CHECKSUM_CORRECT (1 << 27) ++#define STATUS_OVERFLOW_ERR (1 << 28) ++ ++#define UTIL_MAGIC_NUM 0xffd8ffe000104a46 ++#define UTIL_DDRC_WA ++ ++/* The following structure is filled by class-pe when the packet ++ * has to be sent to util-pe, by filling the required information */ ++typedef struct { ++ u32 mtd_flags : 16; ++ u32 packet_type : 8; ++ u32 input_port : 4; ++ u32 data_offset : 4; ++ u32 word[MTD_PRIV]; ++#ifdef UTIL_DDRC_WA ++ u64 magic_num; // magic_number to verify the data validity in utilpe ++#endif ++} __attribute__((aligned(8))) util_rx_hdr_t; // Size must be a multiple of 64-bit to allow copies using EFET. ++ ++#define UTIL_RX_IPS_IN_PKT EVENT_IPS_IN ++#define UTIL_RX_IPS_OUT_PKT EVENT_IPS_OUT ++#define UTIL_RX_RTP_PKT EVENT_RTP_RELAY ++#define UTIL_RX_RTP_QOS_PKT EVENT_RTP_QOS ++#define UTIL_RX_FRAG4_PKT EVENT_FRAG4 ++#define UTIL_RX_FRAG6_PKT EVENT_FRAG6 ++ ++/** Structure passed from UtilPE to Class, stored at the end of the LMEM buffer. Defined and used by software only. ++ * ++ */ ++ ++typedef struct ++{ ++ void *next; ++ u16 next_length; ++ u8 next_l3offset; ++ u8 next_l4offset; ++} frag_info; ++ ++typedef struct { ++ u8 packet_type : 6; ++ u8 padding : 2; ++ ++ u8 offset : 3; ++ u8 ddr_offset : 5; ++ ++ u16 mtd_flags; ++ union { ++ u16 half[6]; ++ u8 byte[12]; ++ ++ struct { ++ u16 sa_handle[2]; // SA_MAX_OP value should be used here instead of 2 ++ u8 proto; ++ S8 sa_op; ++ u8 l2hdr_len; ++ u8 adj_dmem; ++ } ipsec; ++ ++ struct { ++ u16 l4offset; ++ u16 socket_id; ++ BOOL update; ++ u8 reserved; ++ u32 payload_diff; ++ } relay; ++ ++ struct { ++ u16 l3offset; ++ u16 l4offset; ++ ++ frag_info frag; ++ } ipv6; ++ ++ struct { ++ u16 l3offset; ++ } ipv4; ++ ++ struct { ++ u32 ddr_addr; ++ u16 length; ++ u8 port; ++ u8 queue; ++ u8 action; ++ } tx; ++ }; ++} lmem_trailer_t; ++ ++/* The following values are defined for packet_type of lmem_trailer_t. ++ * These represent different types of packets sent from util to class ++ * for processing */ ++enum { ++ UTIL_TX_IPS_IN = 0, ++ UTIL_TX_IPV4_RTP_PKT, ++ UTIL_TX_IPV6_RTP_PKT, ++ UTIL_TX_IPV4_PKT, ++ UTIL_TX_IPV6_PKT, ++ UTIL_TX_EXPT_PKT, ++#ifdef CFG_PCAP ++ UTIL_TX_PKT, ++#endif ++ UTIL_TX_MAX_PKT ++}; ++ ++ ++#define UTIL_TX_TRAILER_SIZE sizeof(lmem_trailer_t) ++#define UTIL_TX_TRAILER(mtd) ((lmem_trailer_t *)ROUND_UP32((u32)(mtd)->rx_dmem_end)) ++ ++typedef struct { ++ u32 pkt_ptr; ++ u8 phyno; ++ u8 queueno; ++ u16 len; ++} tmu_tx_hdr_t; ++ ++struct hif_pkt_hdr { ++ u8 client_id; ++ u8 qNo; ++ u16 client_ctrl_le_lsw; ++ u16 client_ctrl_le_msw; ++}; ++ ++ ++#if defined(CFG_WIFI_OFFLOAD) ++#define MAX_WIFI_VAPS 3 ++#define PFE_WIFI_PKT_HEADROOM 96 /*PFE inserts this headroom for WiFi tx packets only in lro mode */ ++#else ++#define MAX_WIFI_VAPS 0 ++#endif ++ ++/* HIF header client id */ ++enum HIF_CLIENT_ID { ++ CLIENT_ID_GEM0 = 0, ++ CLIENT_ID_GEM1, ++ CLIENT_ID_GEM2, ++ CLIENT_ID_WIFI0, ++ CLIENT_ID_WIFI_LAST = MAX_WIFI_VAPS + CLIENT_ID_GEM2, ++ CLIENT_ID_PCAP, ++ CLIENT_ID_UNKNOWN = 0xff, ++}; ++ ++ ++#define IS_WIFI_CLIENT_ID(_clid) (((_clid) >= CLIENT_ID_WIFI0) && ((_clid) <= CLIENT_ID_WIFI_LAST)) ++ ++/* These match LE definition */ ++#define HIF_CTRL_TX_TSO_NOCPY __cpu_to_le32(1 << 8) ++#define HIF_CTRL_TX_IPSEC_OUT __cpu_to_le32(1 << 7) ++#define HIF_CTRL_TX_WIFI_OWNMAC __cpu_to_le32(1 << 6) ++#define HIF_CTRL_TX_TSO_END __cpu_to_le32(1 << 5) ++#define HIF_CTRL_TX_TSO6 __cpu_to_le32(1 << 4) ++#define HIF_CTRL_TX_TSO __cpu_to_le32(1 << 3) ++#define HIF_CTRL_TX_CHECKSUM __cpu_to_le32(1 << 2) ++#define HIF_CTRL_TX_CSUM_VALIDATE __cpu_to_le32(1 << 1) ++#define HIF_CTRL_TX_WIFI_TXOFLD __cpu_to_le32(1 << 0) ++ ++#define HIF_CTRL_RX_OFFSET_MASK __cpu_to_le32(0xf << 24) ++#define HIF_CTRL_RX_PE_ID_MASK __cpu_to_le32(0xf << 16) ++#define HIF_CTRL_RX_IPSEC_IN __cpu_to_le32(1 << 4) ++#define HIF_CTRL_RX_WIFI_EXPT __cpu_to_le32(1 << 3) ++#define HIF_CTRL_RX_CHECKSUMMED __cpu_to_le32(1 << 2) ++#define HIF_CTRL_RX_CONTINUED __cpu_to_le32(1 << 1) ++#define HIF_CTRL_RX_WIFI_HEADROOM __cpu_to_le32(1 << 0) ++ ++#ifdef CFG_LRO ++struct hif_lro_hdr { ++ u16 data_offset; ++ u16 mss; ++}; ++#endif ++ ++struct hif_ipsec_hdr { ++ u16 sa_handle[2]; ++}; ++ ++#define MAX_TSO_BUF_DESCS 5 ++struct hif_tso_buf_desc { ++ u32 addr; ++ u32 ctrl; ++#define TSO_CTRL_LAST_BUFFER (1 << 31) ++}; ++ ++struct hif_tso_hdr { ++ u16 ip_off; ++ u16 ip_id; ++ u16 ip_len; ++ u16 tcp_off; ++ u32 tcp_seq; ++}; ++ ++struct hif_tso_hdr_nocpy { ++ u16 ip_off; ++ u16 ip_id; ++ u16 ip_len; ++ u16 tcp_off; ++ u32 tcp_seq; ++ struct hif_tso_buf_desc bdesc[MAX_TSO_BUF_DESCS]; ++}; ++ ++struct hif_pcap_hdr { ++ u8 ifindex; ++ u8 unused; ++ u16 seqno; ++ u32 timestamp; ++}; ++ ++ ++struct pe_sync_mailbox ++{ ++ u32 stop; ++ u32 stopped; ++}; ++ ++struct pe_msg_mailbox ++{ ++ u32 dst; ++ u32 src; ++ u32 len; ++ u32 request; ++}; ++ ++ ++/** Basic busy loop delay function ++* ++* @param cycles Number of cycles to delay (actual cpu cycles should be close to 3 x cycles) ++* ++*/ ++static inline void delay(u32 cycles) ++{ ++ volatile int i; ++ ++ for (i = 0; i < cycles; i++); ++} ++ ++ ++/** Read PE id ++* ++* @return PE id (0 - 5 for CLASS-PE's, 6 - 9 for TMU-PE's, 10 for UTIL-PE) ++* ++*/ ++static inline u32 esi_get_mpid(void) ++{ ++ u32 mpid; ++ ++ asm ("rcsr %0, Configuration, MPID" : "=d" (mpid)); ++ ++ return mpid; ++} ++ ++ ++#define esi_get_csr(bank, csr) \ ++({ \ ++ u32 res; \ ++ asm ("rcsr %0, " #bank ", " #csr : "=d" (res)); \ ++ res; \ ++}) ++ ++#define esi_get_isa0() esi_get_csr(Configuration, ISA0) ++#define esi_get_isa1() esi_get_csr(Configuration, ISA1) ++#define esi_get_isa2() esi_get_csr(Configuration, ISA2) ++#define esi_get_isa3() esi_get_csr(Configuration, ISA3) ++#define esi_get_epc() esi_get_csr(Thread, EPC) ++#define esi_get_ecas() esi_get_csr(Thread, ECAS) ++#define esi_get_eid() esi_get_csr(Thread, EID) ++#define esi_get_ed() esi_get_csr(Thread, ED) ++ ++static inline void esi_pe_stop(U32 state) ++{ ++ PESTATUS_SETSTATE(state); ++ while (1) ++ { ++ asm("stop"); ++ } ++} ++ ++ ++/** Same 64bit alignment memory copy using efet. ++* Either the source or destination address must be in DMEM, the other address can be in LMEM or DDR. ++* Both the source and destination must have the same 64bit alignment, length should be more than four bytes ++* or dst/src must be 32bit aligned. Otherwise use efet_memcpy_any() ++* Uses efet synchronous interface to copy the data. ++* ++* @param dst Destination address to write to (must have the same 64bit alignment as src) ++* @param src Source address to read from (must have the same 64bit alignment as dst) ++* @param len Number of bytes to copy ++* ++*/ ++void efet_memcpy(void *dst, void *src, unsigned int len); ++ ++/** Same 64bit alignment memory copy using efet. ++* Either the source or destination address must be in DMEM, the other address can be in LMEM or DDR. ++* Both the source and destination must have the same 64bit alignment, there is no restriction on length. ++* For UTIL-PE revA0, this function will still fail to handle small/unaligned writes. ++* Uses efet synchronous interface to copy the data. ++* ++* @param dst Destination address to write to (must have the same 64bit alignment as src) ++* @param src Source address to read from (must have the same 64bit alignment as dst) ++* @param len Number of bytes to copy ++* ++*/ ++void efet_memcpy_any(void *dst, void *src, unsigned int len); ++ ++/** Same 64bit alignment memory copy using efet. ++* Either the source or destination address must be in DMEM, the other address can be in LMEM or DDR. ++* Both the source and destination must have the same 64bit alignment, length should be more than four bytes ++* or dst/src must be 32bit aligned. ++* Uses efet asynchronous interface to copy the data. ++* ++* @param dst Destination address to write to (must have the same 64bit alignment as src) ++* @param src Source address to read from (must have the same 64bit alignment as dst) ++* @param len Number of bytes to copy ++* ++*/ ++void efet_memcpy_nowait(void *dst, void *src, unsigned int len); ++ ++/** Unaligned memory copy using efet. ++* Either the source or destination address must be in DMEM, the other address can be in LMEM or DDR. ++* There is not restriction on source and destination, nor on length. ++* ++* @param dst Destination address to write to ++* @param src Source address to read from ++* @param len Number of bytes to copy ++* @param dmem_buf temp dmem buffer to use, must be 64bit aligned ++* @param dmem_len length of dmem buffer, must be 64bit aligned and at least 16 bytes ++* ++*/ ++void efet_memcpy_unaligned(void *dst, void *src, unsigned int len, void *dmem_buf, unsigned int dmem_len); ++ ++/** Aligned memory copy of 4 bytes to register address. ++* Register address must be 32 bit aligned. ++* ++* @param val value to be copied. ++* @param reg_addr Register address (must be 16bit aligned) ++* ++*/ ++void __efet_writel(u32 val, void *addr); ++ ++#ifdef REVA_WA ++#define efet_writel(val, addr) __efet_writel((u32)(val), (void *) (addr)) ++#else ++#define efet_writel(val, addr) writel((u32)(val), (void *) (addr)) ++#endif ++ ++ ++/** 32bit aligned memory copy. ++* Source and destination addresses must be 32bit aligned, there is no restriction on the length. ++* ++* @param dst Destination address (must be 32bit aligned) ++* @param src Source address (must be 32bit aligned) ++* @param len Number of bytes to copy ++* ++*/ ++void memcpy_aligned32(void *dst, void *src, unsigned int len); ++ ++/** Aligned memory copy. ++* Source and destination addresses must have the same alignment ++* relative to 32bit boundaries (but otherwsie may have any alignment), ++* there is no restriction on the length. ++* ++* @param dst Destination address ++* @param src Source address (must have same 32bit alignment as dst) ++* @param len Number of bytes to copy ++* ++*/ ++void memcpy_aligned(void *dst, void *src, unsigned int len); ++ ++/** Unaligned memory copy. ++* Implements unaligned memory copy. We first align the destination ++* to a 32bit boundary (using byte copies) then the src, and finally use a loop ++* of read, shift, write ++* ++* @param dst Destination address ++* @param src Source address (must have same 32bit alignment as dst) ++* @param len Number of bytes to copy ++* ++*/ ++void memcpy_unaligned(void *dst, void *src, unsigned int len); ++ ++/** Generic memory set. ++* Implements a generic memory set. Not very optimal (uses byte writes for the entire range) ++* ++* ++* @param dst Destination address ++* @param val Value to set memory to ++* @param len Number of bytes to set ++* ++*/ ++void memset(void *dst, u8 val, unsigned int len); ++ ++/** Generic memory copy. ++* Implements generic memory copy. If source and destination have the same ++* alignment memcpy_aligned() is used, otherwise memcpy_unaligned() ++* ++* @param dst Destination address ++* @param src Source address ++* @param len Number of bytes to copy ++* ++*/ ++void memcpy(void *dst, void *src, unsigned int len); ++ ++/** Generic memorymove. ++* Implements generic memorymove, where copies across overlapping ++* memory regions is supported. ++* Uses the dmem_buf passed as a parameter as a temporary buffer. ++* Includes two copies, forces one of the copies to be definitely aligned. ++* The "dmem_len" being passed should be atleast 3 bytes greater than "len" ++* The 3 bytes here are shift bytes used to ensure one aligned copy. ++* ++* @param dst Destination address ++* @param src Source address ++* @param len Number of bytes to copy ++* @param dmem_buf temp dmem buffer to use, must be 32bit aligned ++* @param dmem_len length of dmem buffer, must be 32bit aligned and at least 3 bytes greater ++* than @param len ++* ++*/ ++ ++void *memorymove(void * dst, void * src, unsigned int len, void *dmem_buf, unsigned int dmem_len); ++ ++/** Aligned memory copy in DDR memory. ++ * Implements aligned memory copy between two DDR buffers using efet_memcpy64 and DMEM ++ * Both the source and destination must have the same 64bit alignment, there is no restriction on length. ++ * If start or end are not 64bit aligned, data in destination buffer before start/after end will be corrupted. ++ * ++ * @param dst DDR Destination address ++ * @param src DDR Source address ++ * @param len Number of bytes to copy ++ * @param dmem_buf temp dmem buffer to use, must be 64bit aligned ++ * @param dmem_len length of dmem buffer, must be 64bit aligned and at least 16 bytes ++ */ ++void memcpy_ddr_to_ddr(void *dst, void *src, unsigned int len, void *dmem_buf, unsigned int dmem_len); ++ ++/** Unaligned memory copy in DDR memory. ++ * Implements generic memory copy between two DDR buffers using efet_memcpy and DMEM ++ * There is no restriction on the source, destination and length alignments. ++ * ++ * @param dst DDR Destination address ++ * @param src DDR Source address ++ * @param len Number of bytes to copy ++ * @param dmem_buf temp dmem buffer to use, must be 64bit aligned ++ * @param dmem_len length of dmem buffer, must be 64bit aligned and at least 16 bytes ++ */ ++void memcpy_ddr_to_ddr_unaligned(void *dst, void *src, unsigned int len, void *dmem_buf, unsigned int dmem_len); ++ ++#endif /* _PE_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/pfe.h +@@ -0,0 +1,444 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _PFE_H_ ++#define _PFE_H_ ++ ++#define CLASS_DMEM_BASE_ADDR(i) (0x00000000 | ((i) << 20)) ++#define CLASS_IMEM_BASE_ADDR(i) (0x00000000 | ((i) << 20)) /* Only valid for mem access register interface */ ++#define CLASS_DMEM_SIZE 0x00002000 ++#define CLASS_IMEM_SIZE 0x00008000 ++ ++#define TMU_DMEM_BASE_ADDR(i) (0x00000000 + ((i) << 20)) ++#define TMU_IMEM_BASE_ADDR(i) (0x00000000 + ((i) << 20)) /* Only valid for mem access register interface */ ++#define TMU_DMEM_SIZE 0x00000800 ++#define TMU_IMEM_SIZE 0x00002000 ++ ++#define UTIL_DMEM_BASE_ADDR 0x00000000 ++#define UTIL_DMEM_SIZE 0x00002000 ++ ++#define PE_LMEM_BASE_ADDR 0xc3010000 ++#define PE_LMEM_SIZE 0x8000 ++#define PE_LMEM_END (PE_LMEM_BASE_ADDR + PE_LMEM_SIZE) ++ ++#define DMEM_BASE_ADDR 0x00000000 ++#define DMEM_SIZE 0x2000 /**< TMU has less... */ ++#define DMEM_END (DMEM_BASE_ADDR + DMEM_SIZE) ++ ++#define PMEM_BASE_ADDR 0x00010000 ++#define PMEM_SIZE 0x8000 /**< TMU has less... */ ++#define PMEM_END (PMEM_BASE_ADDR + PMEM_SIZE) ++ ++ ++/* These check memory ranges from PE point of view/memory map */ ++#define IS_DMEM(addr, len) (((unsigned long)(addr) >= DMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= DMEM_END)) ++#define IS_PMEM(addr, len) (((unsigned long)(addr) >= PMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= PMEM_END)) ++#define IS_PE_LMEM(addr, len) (((unsigned long)(addr) >= PE_LMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= PE_LMEM_END)) ++ ++#define IS_PFE_LMEM(addr, len) (((unsigned long)(addr) >= CBUS_VIRT_TO_PFE(LMEM_BASE_ADDR)) && (((unsigned long)(addr) + (len)) <= CBUS_VIRT_TO_PFE(LMEM_END))) ++#define __IS_PHYS_DDR(addr, len) (((unsigned long)(addr) >= DDR_PHYS_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= DDR_PHYS_END)) ++#define IS_PHYS_DDR(addr, len) __IS_PHYS_DDR(DDR_PFE_TO_PHYS(addr), len) ++ ++/* If using a run-time virtual address for the cbus base address use this code */ ++extern void *cbus_base_addr; ++extern void *ddr_base_addr; ++extern unsigned long ddr_phys_base_addr; ++extern unsigned int ddr_size; ++ ++#if defined(COMCERTO_2000_CONTROL) ++#include <linux/version.h> ++#if defined (CONFIG_PLATFORM_C2000) ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,6,0) ++/*This is copied from arch/arm/include/asm/system_info.h */ ++extern unsigned int system_rev; ++#endif ++#endif ++#endif ++ ++#define CBUS_BASE_ADDR cbus_base_addr ++#define DDR_PHYS_BASE_ADDR ddr_phys_base_addr ++#define DDR_BASE_ADDR ddr_base_addr ++#define DDR_SIZE ddr_size ++ ++#define DDR_PHYS_END (DDR_PHYS_BASE_ADDR + DDR_SIZE) ++ ++#if defined(CONFIG_PLATFORM_C2000) ++#define PFE_CBUS_PHYS_BASE_ADDR 0xc0000000 /**< CBUS physical base address as seen by PE's. */ ++#define DDR_PHYS_TO_PFE(p) (p) ++#define DDR_PFE_TO_PHYS(p) (p) ++#define CBUS_PHYS_TO_PFE(p) (p) ++#else ++#define LS1012A_PFE_RESET_WA /*PFE doesn't have global reset and re-init should takecare few things to make PFE functional after reset */ ++#define PFE_CBUS_PHYS_BASE_ADDR 0xc0000000 /**< CBUS physical base address as seen by PE's. */ ++#define PFE_CBUS_PHYS_BASE_ADDR_FROM_PFE 0xc0000000 /**< CBUS physical base address as seen by PE's. */ ++#define DDR_PHYS_TO_PFE(p) (((unsigned long int) (p)) & 0x7FFFFFFF) ++#define DDR_PFE_TO_PHYS(p) (((unsigned long int) (p)) | 0x80000000) ++#define CBUS_PHYS_TO_PFE(p) (((p) - PFE_CBUS_PHYS_BASE_ADDR) + PFE_CBUS_PHYS_BASE_ADDR_FROM_PFE) /*Translates to PFE address map */ ++#endif ++ ++#define DDR_PHYS_TO_VIRT(p) (((p) - DDR_PHYS_BASE_ADDR) + DDR_BASE_ADDR) ++#define DDR_VIRT_TO_PHYS(v) (((v) - DDR_BASE_ADDR) + DDR_PHYS_BASE_ADDR) ++#define DDR_VIRT_TO_PFE(p) (DDR_PHYS_TO_PFE(DDR_VIRT_TO_PHYS(p))) ++ ++#define CBUS_VIRT_TO_PFE(v) (((v) - CBUS_BASE_ADDR) + PFE_CBUS_PHYS_BASE_ADDR) ++#define CBUS_PFE_TO_VIRT(p) (((p) - PFE_CBUS_PHYS_BASE_ADDR) + CBUS_BASE_ADDR) ++ ++/* The below part of the code is used in QOS control driver from host */ ++#define TMU_APB_BASE_ADDR 0xc1000000 /** TMU base address seen by pe's */ ++ ++#define SHAPER0_BASE_ADDR (TMU_APB_BASE_ADDR + 0x020000) ++#define SHAPER1_BASE_ADDR (TMU_APB_BASE_ADDR + 0x030000) ++#define SHAPER2_BASE_ADDR (TMU_APB_BASE_ADDR + 0x040000) ++#define SHAPER3_BASE_ADDR (TMU_APB_BASE_ADDR + 0x050000) ++#define SHAPER4_BASE_ADDR (TMU_APB_BASE_ADDR + 0x060000) ++#define SHAPER5_BASE_ADDR (TMU_APB_BASE_ADDR + 0x070000) ++#define SHAPER6_BASE_ADDR (TMU_APB_BASE_ADDR + 0x080000) ++#define SHAPER7_BASE_ADDR (TMU_APB_BASE_ADDR + 0x090000) ++#define SHAPER8_BASE_ADDR (TMU_APB_BASE_ADDR + 0x0a0000) ++#define SHAPER9_BASE_ADDR (TMU_APB_BASE_ADDR + 0x0b0000) ++ ++#define SCHED0_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1c0000) ++#define SCHED1_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1d0000) ++#define SCHED2_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1e0000) ++#define SCHED3_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1f0000) ++#define SCHED4_BASE_ADDR (TMU_APB_BASE_ADDR + 0x200000) ++#define SCHED5_BASE_ADDR (TMU_APB_BASE_ADDR + 0x210000) ++#define SCHED6_BASE_ADDR (TMU_APB_BASE_ADDR + 0x220000) ++#define SCHED7_BASE_ADDR (TMU_APB_BASE_ADDR + 0x230000) ++ ++#define PHY_QUEUE_BASE_ADDR (TMU_APB_BASE_ADDR + 0x260000) ++#define QUEUE_RESULT0 (PHY_QUEUE_BASE_ADDR + 0x48) /**< [7] set to one to indicate output PHY (TMU0->PHY0, TMU1->PHY1, TMU2->PHY2, TMU3->PHY3), [6:0] winner input queue number */ ++#define QUEUE_RESULT1 (PHY_QUEUE_BASE_ADDR + 0x4c) /**< [7] set to one to indicate output PHY (TMU0->PHY0, TMU1->PHY1, TMU2->PHY2, TMU3->PHY4), [6:0] winner input queue number */ ++#define QUEUE_RESULT2 (PHY_QUEUE_BASE_ADDR + 0x50) /**< [7] set to one to indicate output PHY (TMU0->PHY0, TMU1->PHY1, TMU2->PHY2, TMU3->PHY5), [6:0] winner input queue number */ ++ ++#define QUEUE_RESULT0_REGOFFSET (QUEUE_RESULT0 - QUEUE_RESULT0) ++#define QUEUE_RESULT1_REGOFFSET (QUEUE_RESULT1 - QUEUE_RESULT0) ++#define QUEUE_RESULT2_REGOFFSET (QUEUE_RESULT2 - QUEUE_RESULT0) ++ ++ ++#include "cbus.h" ++ ++enum { ++ CLASS0_ID = 0, ++ CLASS1_ID, ++ CLASS2_ID, ++ CLASS3_ID, ++#if !defined(CONFIG_PLATFORM_PCI) ++ CLASS4_ID, ++ CLASS5_ID, ++#endif ++#if !defined(CONFIG_TMU_DUMMY) ++ TMU0_ID, ++ TMU1_ID, ++ TMU2_ID, ++ TMU3_ID, ++#else ++ TMU0_ID, ++#endif ++#if !defined(CONFIG_UTIL_DISABLED) ++ UTIL_ID, ++#endif ++ MAX_PE ++}; ++ ++enum { ++ CLASS_TYPE = 0, ++ TMU_TYPE, ++ UTIL_TYPE ++}; ++ ++#if !defined(CONFIG_PLATFORM_PCI) ++#define CLASS_MASK ((1 << CLASS0_ID) | (1 << CLASS1_ID) | (1 << CLASS2_ID) | (1 << CLASS3_ID) | (1 << CLASS4_ID) | (1 << CLASS5_ID)) ++#define CLASS_MAX_ID CLASS5_ID ++#else ++#define CLASS_MASK ((1 << CLASS0_ID) | (1 << CLASS1_ID) | (1 << CLASS2_ID) | (1 << CLASS3_ID)) ++#define CLASS_MAX_ID CLASS3_ID ++#endif ++ ++#if !defined(CONFIG_TMU_DUMMY) ++#if defined(CONFIG_PLATFORM_LS1012A) ++#define TMU_MASK ((1 << TMU0_ID) | (1 << TMU1_ID) | (1 << TMU3_ID)) ++#else ++#define TMU_MASK ((1 << TMU0_ID) | (1 << TMU1_ID) | (1 << TMU2_ID) | (1 << TMU3_ID)) ++#endif ++#define TMU_MAX_ID TMU3_ID ++#else ++#define TMU_MASK (1 << TMU0_ID) ++#define TMU_MAX_ID TMU0_ID ++#endif ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++#define UTIL_MASK (1 << UTIL_ID) ++#endif ++ ++typedef struct tPE_STATUS ++{ ++ u32 cpu_state; ++ u32 activity_counter; ++ u32 rx; ++ union { ++ u32 tx; ++ u32 tmu_qstatus; ++ }; ++ u32 drop; ++#if defined(CFG_PE_DEBUG) ++ u32 debug_indicator; ++ u32 debug[16]; ++#endif ++} __attribute__((aligned(16))) PE_STATUS; ++ ++ ++struct pe_sync_mailbox ++{ ++ u32 stop; ++ u32 stopped; ++}; ++ ++struct pe_msg_mailbox ++{ ++ u32 dst; ++ u32 src; ++ u32 len; ++ u32 request; ++}; ++ ++// Drop counter definitions ++ ++#define CLASS_NUM_DROP_COUNTERS 13 ++#define UTIL_NUM_DROP_COUNTERS 8 ++ ++ ++/** PE information. ++ * Structure containing PE's specific information. It is used to create ++ * generic C functions common to all PE's. ++ * Before using the library functions this structure needs to be initialized with the different registers virtual addresses ++ * (according to the ARM MMU mmaping). The default initialization supports a virtual == physical mapping. ++ * ++ */ ++struct pe_info ++{ ++ u32 dmem_base_addr; /**< PE's dmem base address */ ++ u32 pmem_base_addr; /**< PE's pmem base address */ ++ u32 pmem_size; /**< PE's pmem size */ ++ ++ void *mem_access_wdata; /**< PE's _MEM_ACCESS_WDATA register address */ ++ void *mem_access_addr; /**< PE's _MEM_ACCESS_ADDR register address */ ++ void *mem_access_rdata; /**< PE's _MEM_ACCESS_RDATA register address */ ++}; ++ ++ ++void pe_lmem_read(u32 *dst, u32 len, u32 offset); ++void pe_lmem_write(u32 *src, u32 len, u32 offset); ++ ++void pe_dmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len); ++void pe_pmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len); ++ ++u32 pe_pmem_read(int id, u32 addr, u8 size); ++ ++void pe_dmem_write(int id, u32 val, u32 addr, u8 size); ++u32 pe_dmem_read(int id, u32 addr, u8 size); ++void class_pe_lmem_memcpy_to32(u32 dst, const void *src, unsigned int len); ++void class_pe_lmem_memset(u32 dst, int val, unsigned int len); ++void class_bus_write(u32 val, u32 addr, u8 size); ++u32 class_bus_read(u32 addr, u8 size); ++ ++ ++#define class_bus_readl(addr) class_bus_read(addr, 4) ++#define class_bus_readw(addr) class_bus_read(addr, 2) ++#define class_bus_readb(addr) class_bus_read(addr, 1) ++ ++#define class_bus_writel(val, addr) class_bus_write(val, addr, 4) ++#define class_bus_writew(val, addr) class_bus_write(val, addr, 2) ++#define class_bus_writeb(val, addr) class_bus_write(val, addr, 1) ++ ++#define pe_dmem_readl(id, addr) pe_dmem_read(id, addr, 4) ++#define pe_dmem_readw(id, addr) pe_dmem_read(id, addr, 2) ++#define pe_dmem_readb(id, addr) pe_dmem_read(id, addr, 1) ++ ++#define pe_dmem_writel(id, val, addr) pe_dmem_write(id, val, addr, 4) ++#define pe_dmem_writew(id, val, addr) pe_dmem_write(id, val, addr, 2) ++#define pe_dmem_writeb(id, val, addr) pe_dmem_write(id, val, addr, 1) ++ ++//int pe_load_elf_section(int id, const void *data, Elf32_Shdr *shdr); ++int pe_load_elf_section(int id, const void *data, Elf32_Shdr *shdr, struct device *dev); ++ ++void pfe_lib_init(void *cbus_base, void *ddr_base, unsigned long ddr_phys_base, unsigned int ddr_size); ++void bmu_init(void *base, BMU_CFG *cfg); ++void bmu_reset(void *base); ++void bmu_enable(void *base); ++void bmu_disable(void *base); ++void bmu_set_config(void *base, BMU_CFG *cfg); ++ ++/* An enumerated type for loopback values. This can be one of three values, no ++ * loopback -normal operation, local loopback with internal loopback module of ++ * MAC or PHY loopback which is through the external PHY. ++ */ ++#ifndef __MAC_LOOP_ENUM__ ++#define __MAC_LOOP_ENUM__ ++typedef enum {LB_NONE, LB_EXT, LB_LOCAL} MAC_LOOP; ++#endif ++ ++ ++void gemac_init(void *base, void *config); ++void gemac_disable_rx_checksum_offload(void *base); ++void gemac_enable_rx_checksum_offload(void *base); ++void gemac_set_mdc_div(void *base, int mdc_div); ++void gemac_set_speed(void *base, MAC_SPEED gem_speed); ++void gemac_set_duplex(void *base, int duplex); ++void gemac_set_mode(void *base, int mode); ++void gemac_enable(void *base); ++void gemac_tx_disable(void *base); ++void gemac_disable(void *base); ++void gemac_reset(void *base); ++void gemac_set_address(void *base, SPEC_ADDR *addr); ++SPEC_ADDR gemac_get_address(void *base); ++void gemac_set_loop( void *base, MAC_LOOP gem_loop ); ++void gemac_set_laddr1(void *base, MAC_ADDR *address); ++void gemac_set_laddr2(void *base, MAC_ADDR *address); ++void gemac_set_laddr3(void *base, MAC_ADDR *address); ++void gemac_set_laddr4(void *base, MAC_ADDR *address); ++void gemac_set_laddrN(void *base, MAC_ADDR *address, unsigned int entry_index); ++void gemac_clear_laddr1(void *base); ++void gemac_clear_laddr2(void *base); ++void gemac_clear_laddr3(void *base); ++void gemac_clear_laddr4(void *base); ++void gemac_clear_laddrN(void *base, unsigned int entry_index); ++MAC_ADDR gemac_get_hash( void *base ); ++void gemac_set_hash( void *base, MAC_ADDR *hash ); ++MAC_ADDR gem_get_laddr1(void *base); ++MAC_ADDR gem_get_laddr2(void *base); ++MAC_ADDR gem_get_laddr3(void *base); ++MAC_ADDR gem_get_laddr4(void *base); ++MAC_ADDR gem_get_laddrN(void *base, unsigned int entry_index); ++void gemac_set_config(void *base, GEMAC_CFG *cfg); ++void gemac_allow_broadcast(void *base); ++void gemac_no_broadcast(void *base); ++void gemac_enable_unicast(void *base); ++void gemac_disable_unicast(void *base); ++void gemac_enable_multicast(void *base); ++void gemac_disable_multicast(void *base); ++void gemac_enable_fcs_rx(void *base); ++void gemac_disable_fcs_rx(void *base); ++void gemac_enable_1536_rx(void *base); ++void gemac_disable_1536_rx(void *base); ++void gemac_enable_rx_jmb(void *base); ++void gemac_disable_rx_jmb(void *base); ++void gemac_enable_stacked_vlan(void *base); ++void gemac_disable_stacked_vlan(void *base); ++void gemac_enable_pause_rx(void *base); ++void gemac_disable_pause_rx(void *base); ++void gemac_enable_copy_all(void *base); ++void gemac_disable_copy_all(void *base); ++void gemac_set_bus_width(void *base, int width); ++void gemac_set_wol(void *base, u32 wol_conf); ++ ++void gpi_init(void *base, GPI_CFG *cfg); ++void gpi_reset(void *base); ++void gpi_enable(void *base); ++void gpi_disable(void *base); ++void gpi_set_config(void *base, GPI_CFG *cfg); ++ ++void class_init(CLASS_CFG *cfg); ++void class_reset(void); ++void class_enable(void); ++void class_disable(void); ++void class_set_config(CLASS_CFG *cfg); ++ ++void tmu_reset(void); ++void tmu_init(TMU_CFG *cfg); ++void tmu_enable(u32 pe_mask); ++void tmu_disable(u32 pe_mask); ++u32 tmu_qstatus(u32 if_id); ++u32 tmu_pkts_processed(u32 if_id); ++ ++void util_init(UTIL_CFG *cfg); ++void util_reset(void); ++void util_enable(void); ++void util_disable(void); ++ ++void hif_nocpy_init(void); ++void hif_nocpy_tx_enable(void); ++void hif_nocpy_tx_disable(void); ++void hif_nocpy_rx_enable(void); ++void hif_nocpy_rx_disable(void); ++ ++void hif_init(void); ++void hif_tx_enable(void); ++void hif_tx_disable(void); ++void hif_rx_enable(void); ++void hif_rx_disable(void); ++ ++ ++/** Get Chip Revision level ++* ++*/ ++ ++static inline unsigned int CHIP_REVISION(void) ++{ ++#if defined (CONFIG_PLATFORM_C2000) ++#if 1 ++ return system_rev; ++ //return 0; ++#else ++ return (readl(COMCERTO_GPIO_DEVICE_ID_REG) >> 24) & 0xf; ++#endif ++ ++#else ++ /*For LS1012A return always 1 */ ++ return 1; ++#endif ++} ++ ++/** Start HIF rx DMA ++* ++*/ ++static inline void hif_rx_dma_start(void) ++{ ++ /*TODO not sure poll_cntrl_en is required or not */ ++ writel(HIF_CTRL_DMA_EN | HIF_CTRL_BDP_CH_START_WSTB, HIF_RX_CTRL); ++} ++ ++/** Start HIF tx DMA ++* ++*/ ++static inline void hif_tx_dma_start(void) ++{ ++ /*TODO not sure poll_cntrl_en is required or not */ ++ writel(HIF_CTRL_DMA_EN | HIF_CTRL_BDP_CH_START_WSTB, HIF_TX_CTRL); ++} ++ ++/** Start HIF_NOCPY rx DMA ++* ++*/ ++static inline void hif_nocpy_rx_dma_start(void) ++{ ++ /*TODO not sure poll_cntrl_en is required or not */ ++ writel((HIF_CTRL_DMA_EN | HIF_CTRL_BDP_CH_START_WSTB), HIF_NOCPY_RX_CTRL); ++} ++ ++/** Start HIF_NOCPY tx DMA ++* ++*/ ++static inline void hif_nocpy_tx_dma_start(void) ++{ ++ /*TODO not sure poll_cntrl_en is required or not */ ++ writel(HIF_CTRL_DMA_EN | HIF_CTRL_BDP_CH_START_WSTB, HIF_NOCPY_TX_CTRL); ++} ++ ++#endif /* _PFE_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/tmu.h +@@ -0,0 +1,68 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _TMU_H_ ++#define _TMU_H_ ++ ++#define TMU_DMEM_BASE_ADDR 0x00000000 ++#define TMU_PMEM_BASE_ADDR 0x00010000 ++ ++#define CBUS_BASE_ADDR 0xc0000000 ++#define TMU_APB_BASE_ADDR 0xc1000000 ++ ++#if defined (COMCERTO_2000_TMU) || defined (COMCERTO_2000_CONTROL) ++ ++#include "cbus.h" ++ ++#define GPT_BASE_ADDR (TMU_APB_BASE_ADDR + 0x00000) ++#define UART_BASE_ADDR (TMU_APB_BASE_ADDR + 0x10000) ++ ++#define SHAPER0_BASE_ADDR (TMU_APB_BASE_ADDR + 0x020000) ++#define SHAPER1_BASE_ADDR (TMU_APB_BASE_ADDR + 0x030000) ++#define SHAPER2_BASE_ADDR (TMU_APB_BASE_ADDR + 0x040000) ++#define SHAPER3_BASE_ADDR (TMU_APB_BASE_ADDR + 0x050000) ++#define SHAPER4_BASE_ADDR (TMU_APB_BASE_ADDR + 0x060000) ++#define SHAPER5_BASE_ADDR (TMU_APB_BASE_ADDR + 0x070000) ++#define SHAPER6_BASE_ADDR (TMU_APB_BASE_ADDR + 0x080000) ++#define SHAPER7_BASE_ADDR (TMU_APB_BASE_ADDR + 0x090000) ++#define SHAPER8_BASE_ADDR (TMU_APB_BASE_ADDR + 0x0a0000) ++#define SHAPER9_BASE_ADDR (TMU_APB_BASE_ADDR + 0x0b0000) ++ ++#define SCHED0_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1c0000) ++#define SCHED1_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1d0000) ++#define SCHED2_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1e0000) ++#define SCHED3_BASE_ADDR (TMU_APB_BASE_ADDR + 0x1f0000) ++#define SCHED4_BASE_ADDR (TMU_APB_BASE_ADDR + 0x200000) ++#define SCHED5_BASE_ADDR (TMU_APB_BASE_ADDR + 0x210000) ++#define SCHED6_BASE_ADDR (TMU_APB_BASE_ADDR + 0x220000) ++#define SCHED7_BASE_ADDR (TMU_APB_BASE_ADDR + 0x230000) ++ ++#define SHAPER_STATUS (TMU_APB_BASE_ADDR + 0x270000) /**< [9:0] bitmask of shapers that have positive credit */ ++ ++#include "gpt.h" ++#include "uart.h" ++#include "tmu/shaper.h" ++#include "tmu/sched.h" ++ ++#endif ++ ++#define PHY_QUEUE_BASE_ADDR (TMU_APB_BASE_ADDR + 0x260000) ++ ++#include "tmu/phy_queue.h" ++ ++#endif /* _TMU_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/tmu/phy_queue.h +@@ -0,0 +1,56 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _PHY_QUEUE_H_ ++#define _PHY_QUEUE_H_ ++ ++#define PHY_QUEUE_SHAPER_STATUS (PHY_QUEUE_BASE_ADDR + 0x00) /**< [28:19] same as SHAPER_STATUS, [18:3] same as QUEUE_STATUS, [2:0] must be zero before a new packet may be dequeued */ ++#define QUEUE_STATUS (PHY_QUEUE_BASE_ADDR + 0x04) /**< [15:0] bit mask of input queues with pending packets */ ++ ++#define QUEUE0_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x08) ++#define QUEUE1_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x0c) ++#define QUEUE2_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x10) ++#define QUEUE3_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x14) ++#define QUEUE4_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x18) ++#define QUEUE5_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x1c) ++#define QUEUE6_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x20) ++#define QUEUE7_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x24) ++#define QUEUE8_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x28) ++#define QUEUE9_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x2c) ++#define QUEUE10_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x30) ++#define QUEUE11_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x34) ++#define QUEUE12_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x38) ++#define QUEUE13_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x3c) ++#define QUEUE14_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x40) ++#define QUEUE15_PKT_LEN (PHY_QUEUE_BASE_ADDR + 0x44) ++#define QUEUE_RESULT0 (PHY_QUEUE_BASE_ADDR + 0x48) /**< [7] set to one to indicate output PHY (TMU0->PHY0, TMU1->PHY1, TMU2->PHY2, TMU3->PHY3), [6:0] winner input queue number */ ++#define QUEUE_RESULT1 (PHY_QUEUE_BASE_ADDR + 0x4c) /**< [7] set to one to indicate output PHY (TMU0->PHY0, TMU1->PHY1, TMU2->PHY2, TMU3->PHY4), [6:0] winner input queue number */ ++#define QUEUE_RESULT2 (PHY_QUEUE_BASE_ADDR + 0x50) /**< [7] set to one to indicate output PHY (TMU0->PHY0, TMU1->PHY1, TMU2->PHY2, TMU3->PHY5), [6:0] winner input queue number */ ++#define TMU_PE_GP_REG (PHY_QUEUE_BASE_ADDR + 0x54) ++#define QUEUE_GBL_PKTLEN (PHY_QUEUE_BASE_ADDR + 0x5c) ++#define QUEUE_GBL_PKTLEN_MASK (PHY_QUEUE_BASE_ADDR + 0x60) ++ ++#define QUEUE_RESULT0_REGOFFSET (QUEUE_RESULT0 - QUEUE_RESULT0) ++#define QUEUE_RESULT1_REGOFFSET (QUEUE_RESULT1 - QUEUE_RESULT0) ++#define QUEUE_RESULT2_REGOFFSET (QUEUE_RESULT2 - QUEUE_RESULT0) ++ ++#define TEQ_HTD (1 << 22) ++#define TEQ_HWRED (1 << 21) ++ ++ ++#endif /* _PHY_QUEUE_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/tmu/sched.h +@@ -0,0 +1,72 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _SCHED_H_ ++#define _SCHED_H_ ++ ++/* Offsets from SCHEDx_BASE_ADDR */ ++#define SCHED_CTRL 0x00 ++#define SCHED_SLOT_TIME 0x04 ++#define SCHED_RES 0x08 ++#define SCHED_QUEUE_ALLOC0 0x0c ++#define SCHED_QUEUE_ALLOC1 0x10 ++#define SCHED_BW 0x14 ++#define SCHED_GUR_DEF_CTR 0x18 ++#define SCHED_AVL_CTR 0x1c ++#define SCHED_QU0_WGHT 0x20 ++#define SCHED_QU1_WGHT 0x24 ++#define SCHED_QU2_WGHT 0x28 ++#define SCHED_QU3_WGHT 0x2c ++#define SCHED_QU4_WGHT 0x30 ++#define SCHED_QU5_WGHT 0x34 ++#define SCHED_QU6_WGHT 0x38 ++#define SCHED_QU7_WGHT 0x3c ++#define SCHED_QUE0_DEFICIT_CNT 0x40 ++#define SCHED_QUE1_DEFICIT_CNT 0x44 ++#define SCHED_QUE2_DEFICIT_CNT 0x48 ++#define SCHED_QUE3_DEFICIT_CNT 0x4c ++#define SCHED_QUE4_DEFICIT_CNT 0x50 ++#define SCHED_QUE5_DEFICIT_CNT 0x54 ++#define SCHED_QUE6_DEFICIT_CNT 0x58 ++#define SCHED_QUE7_DEFICIT_CNT 0x5c ++#define SCHED_PKT_LEN 0x60 ++ ++#define SCHED_CTRL_ALGOTYPE(x) (((x) & 0xf) << 0) ++#define SCHED_CTRL_CALQUOTA(x) (((x) & 0x1) << 4) ++#define SCHED_CTRL_ACTIVE_Q(x) (((x) & 0xff) << 8) ++#define SCHED_CTRL_SHARE_BW(x) (((x) & 0xff) << 16) ++#define SCHED_CTRL_BARROW_BW(x) (((x) & 0xff) << 24) ++ ++#define SCHED_QUEUE_ALLOC(x, b) (((x) & 0x1f) << (b)) ++ ++#define SCHED_QUEUE_ALLOC0_QUEUEA(x) (((x) & 0x1f) << 0) ++#define SCHED_QUEUE_ALLOC0_QUEUEB(x) (((x) & 0x1f) << 8) ++#define SCHED_QUEUE_ALLOC0_QUEUEC(x) (((x) & 0x1f) << 16) ++#define SCHED_QUEUE_ALLOC0_QUEUED(x) (((x) & 0x1f) << 24) ++ ++#define SCHED_QUEUE_ALLOC0_RES0(x) (((x) & 0x7) << 5) ++#define SCHED_QUEUE_ALLOC0_RES1(x) (((x) & 0x7) << 13) ++#define SCHED_QUEUE_ALLOC0_RES2(x) (((x) & 0x7) << 21) ++#define SCHED_QUEUE_ALLOC0_RES3(x) (((x) & 0x7) << 29) ++ ++#define SCHED_QUEUE_ALLOC1_QUEUEA(x) (((x) & 0x1f) << 0) ++#define SCHED_QUEUE_ALLOC1_QUEUEB(x) (((x) & 0x1f) << 8) ++#define SCHED_QUEUE_ALLOC1_QUEUEC(x) (((x) & 0x1f) << 16) ++#define SCHED_QUEUE_ALLOC1_QUEUED(x) (((x) & 0x1f) << 24) ++ ++#endif /* _SCHED_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/tmu/shaper.h +@@ -0,0 +1,37 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _SHAPER_H_ ++#define _SHAPER_H_ ++ ++/* Offsets from SHAPPERx_BASE_ADDR */ ++#define SHAPER_CTRL 0x00 ++#define SHAPER_WEIGHT 0x04 ++#define SHAPER_PKT_LEN 0x08 ++ ++#define SHAPER_CTRL_ENABLE(x) (((x) & 0x1) << 0) ++#define SHAPER_CTRL_QNO(x) (((x) & 0x3f) << 1) ++#define SHAPER_CTRL_CLKDIV(x) (((x) & 0xffff) << 16) ++ ++#define SHAPER_WEIGHT_FRACWT(x) (((x) & 0xff) << 0) ++#define SHAPER_WEIGHT_INTWT(x) (((x) & 0x3) << 8) ++#define SHAPER_WEIGHT_MAXCREDIT(x) (((x) & 0x3fffff) << 10) ++ ++#define PORT_SHAPER_MASK (1 << 0) ++ ++#endif /* _SHAPER_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/uart.h +@@ -0,0 +1,31 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _UART_H_ ++#define _UART_H_ ++ ++#define UART_THR (UART_BASE_ADDR + 0x00) ++#define UART_IER (UART_BASE_ADDR + 0x04) ++#define UART_IIR (UART_BASE_ADDR + 0x08) ++#define UART_LCR (UART_BASE_ADDR + 0x0c) ++#define UART_MCR (UART_BASE_ADDR + 0x10) ++#define UART_LSR (UART_BASE_ADDR + 0x14) ++#define UART_MDR (UART_BASE_ADDR + 0x18) ++#define UART_SCRATCH (UART_BASE_ADDR + 0x1c) ++ ++#endif /* _UART_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/util.h +@@ -0,0 +1,49 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _UTIL_H_ ++#define _UTIL_H_ ++ ++#define UTIL_DMEM_BASE_ADDR 0x00000000 ++#define UTIL_DMEM_SIZE 0x00002000 ++#define UTIL_DMEM_END (UTIL_DMEM_BASE_ADDR + UTIL_DMEM_SIZE) ++ ++#define IS_DMEM(addr, len) (((unsigned long)(addr) >= UTIL_DMEM_BASE_ADDR) && (((unsigned long)(addr) + (len)) <= UTIL_DMEM_END)) ++ ++#define CBUS_BASE_ADDR 0xc0000000 ++#define UTIL_APB_BASE_ADDR 0xc1000000 ++ ++#include "cbus.h" ++ ++#define GPT_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x00000) ++#define UART_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x10000) ++#define EAPE_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x20000) ++#define INQ_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x30000) ++#define EFET1_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x40000) ++#define EFET2_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x50000) ++#define EFET3_BASE_ADDR (UTIL_APB_BASE_ADDR + 0x60000) ++ ++ ++#include "gpt.h" ++#include "uart.h" ++#include "util/eape.h" ++#include "util/inq.h" ++#include "util/efet.h" ++ ++ ++#endif /* _UTIL_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/util/eape.h +@@ -0,0 +1,57 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _EAPE_H_ ++#define _EAPE_H_ ++ ++#define EAPE_STATUS (EAPE_BASE_ADDR + 0x0) ++#define EAPE_INT_ENABLE (EAPE_BASE_ADDR + 0x4) ++#define EAPE_INT_SRC (EAPE_BASE_ADDR + 0x8) ++#define EAPE_HOST_INT_ENABLE (EAPE_BASE_ADDR + 0xc) ++ ++/** The following bits represents to enable interrupts from host and to host ++* from / to utilpe */ ++ ++#define IRQ_EN_EFET_TO_UTIL 0x1 ++#define IRQ_EN_QB_TO_UTIL 0x2 ++#define IRQ_EN_INQ_TO_UTIL 0x4 ++#define IRQ_EN_EAPE_TO_UTIL 0x8 ++#define IRQ_EN_GPT_TMR_TO_UTIL 0x10 ++#define IRQ_EN_UART_TO_UTIL 0x20 ++#define IRQ_EN_SYSLP_TO_UTIL 0x40 ++#define IRQ_EN_UPEGP_TO_UTIL 0x80 ++ ++/** Out interrupts */ ++ ++#define IRQ_EN_EFET_OUT 0x100 ++#define IRQ_EN_QB_OUT 0x200 ++#define IRQ_EN_INQ_OUT 0x400 ++#define IRQ_EN_EAPE_OUT 0x800 ++#define IRQ_EN_GPT_TMR_OUT 0x1000 ++#define IRQ_EN_UART_OUT 0x2000 ++#define IRQ_EN_SYSLP_OUT 0x4000 ++#define IRQ_EN_UPEGP_OUT 0x8000 ++ ++/** The following bits are enabled in the status register ++ * which are mapped to IPSEC status register bits */ ++#define EAPE_IN_STAT_AVAIL 0x1 ++#define EAPE_OUT_STAT_AVAIL 0x2 ++#define EAPE_IN_CMD_AVAIL 0x4 ++#define EAPE_OUT_CMD_AVAIL 0x8 ++ ++#endif /* _EAPE_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/util/efet.h +@@ -0,0 +1,119 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _UTIL_EFET_H_ ++#define _UTIL_EFET_H_ ++ ++#define EFET_ENTRY_ADDR 0x00 ++#define EFET_ENTRY_SIZE 0x04 ++#define EFET_ENTRY_DMEM_ADDR 0x08 ++#define EFET_ENTRY_STATUS 0x0c ++#define EFET_ENTRY_ENDIAN 0x10 ++ ++#define CBUS2DMEM 0 ++#define DMEM2CBUS 1 ++ ++#define EFET2BUS_LE (1 << 0) ++ ++#define EFET1 0 ++#define EFET2 1 ++#define EFET3 2 ++#define MAX_UTIL_EFET_LEN 128 ++ ++extern const unsigned long util_efet_baseaddr[3]; ++extern u32 util_efet_status; ++ ++/* The barrier call is an empirical work-around for an unknown bug: for some unknown reason, it solves ++ * a UtilPE crash observed with LRO and packet steering. Other solutions also worked (e.g. barrier, ++ * nop calls in other positions). However, no common pattern could be extracted from those solutions ++ * to narrow down the source of the crash. ++ */ ++ ++#define __UTIL_EFET(i, cbus_addr, dmem_addr,len,dir) do { \ ++ __writel((len & 0x3FF) | (dir << 16), util_efet_baseaddr[i] + EFET_ENTRY_SIZE); \ ++ __writel(dmem_addr, util_efet_baseaddr[i] + EFET_ENTRY_DMEM_ADDR);\ ++ __writel(cbus_addr, util_efet_baseaddr[i] + EFET_ENTRY_ADDR);\ ++ nop();\ ++ }while(0) ++ ++#define UTIL_EFET(i, cbus_addr, dmem_addr,len,dir) do { \ ++ __UTIL_EFET(i, cbus_addr, dmem_addr, len, dir); \ ++ util_efet_status |= (1 << i); \ ++ } while(0) ++ ++ ++/** Waits for the util efet to finish a transaction, blocking the caller ++* (without updating the status). ++* Can be called at any time. ++* ++* @param i Efet index ++* ++* ++*/ ++static inline void __util_efet_wait(int i) ++{ ++ while (!(readl(util_efet_baseaddr[i] + EFET_ENTRY_STATUS) & 0x1)) ; ++} ++ ++/** Waits for the util efet to finish a transaction, blocking the caller. ++* Can be called at any time. ++* ++* @param i Efet index ++* ++*/ ++static inline void util_efet_wait(int i) ++{ ++ __util_efet_wait(i); ++ ++ util_efet_status &= ~(1 << i); ++} ++ ++/** Asynchronous interface to util efet read/write functions. ++* It will wait for the efet to finish previous transaction, but does not wait for the current transaction to finish. ++* ++* @param i Efet index ++* @param cbus_addr Cbus address (must be 64bits aligned) ++* @param dmem_addr DMEM address (must be 64bits aligned) ++* @param len Number of bytes to copy (must be 64bits aligned size) ++* @param dir Direction of the transaction (0 - cbus to dmem, 1 - dmem to cbus) ++* ++*/ ++static inline void util_efet_async(int i, u32 cbus_addr, u32 dmem_addr, u32 len, u8 dir) ++{ ++ if (util_efet_status & (1 << i)) ++ util_efet_wait(i); ++ ++ UTIL_EFET(i, cbus_addr, dmem_addr, len, dir); ++} ++ ++ ++static inline void util_efet_async0( u32 cbus_addr, u32 dmem_addr, u32 len, u8 dir) ++{ ++ util_efet_async(0, cbus_addr, dmem_addr, len,dir); ++} ++ ++/* EFET 2 is aways used for SYNC operations */ ++static inline void util_efet_sync2(u32 cbus_addr, u32 dmem_addr, u32 len, u8 dir) ++{ ++ __UTIL_EFET(2, cbus_addr, dmem_addr, len,dir); ++ __util_efet_wait(2); ++} ++ ++void util_efet_sync0(u32 cbus_addr, u32 dmem_addr, u32 len, u8 dir); ++#endif /* _UTIL_EFET_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/include/pfe/util/inq.h +@@ -0,0 +1,28 @@ ++/* ++ * Copyright (c) 2011, 2014 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License ++ * as published by the Free Software Foundation; either version 2 ++ * of the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. ++ * ++*/ ++#ifndef _INQ_H_ ++#define _INQ_H_ ++ ++#define INQ_HOST_GP (INQ_BASE_ADDR + 0x00) /* FIXME what are these for ? */ ++#define INQ_UPE_GP (INQ_BASE_ADDR + 0x04) /* FIXME what are these for ? */ ++ ++#define INQ_QB_PKTPTR (INQ_BASE_ADDR + 0x08) ++#define INQ_FIFO_CNT (INQ_BASE_ADDR + 0x0c) ++ ++#endif /* _INQ_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_ctrl.c +@@ -0,0 +1,363 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifdef __KERNEL__ ++#include <linux/kernel.h> ++#include <linux/sched.h> ++#include <linux/module.h> ++#include <linux/list.h> ++#include <linux/kthread.h> ++#else ++#include "platform.h" ++#endif ++ ++#include "pfe_mod.h" ++#include "pfe_ctrl.h" ++ ++#include "pfe_ctrl_hal.h" ++ ++static struct pe_sync_mailbox CLASS_DMEM_SH2(sync_mailbox); ++static struct pe_sync_mailbox TMU_DMEM_SH2(sync_mailbox); ++ ++static struct pe_msg_mailbox CLASS_DMEM_SH2(msg_mailbox); ++static struct pe_msg_mailbox TMU_DMEM_SH2(msg_mailbox); ++ ++#if !defined(CONFIG_PLATFORM_LS1012A) ++static u32 CLASS_DMEM_SH2(resume); ++static u32 TMU_DMEM_SH2(resume); ++#endif ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++static struct pe_sync_mailbox UTIL_DMEM_SH2(sync_mailbox); ++static struct pe_msg_mailbox UTIL_DMEM_SH2(msg_mailbox); ++static u32 UTIL_DMEM_SH2(resume); ++#endif ++ ++static int pfe_ctrl_timer(void *data); ++ ++static int initialized = 0; ++ ++#define TIMEOUT_MS 1000 ++ ++int relax(unsigned long end) ++{ ++#ifdef __KERNEL__ ++ if (time_after(jiffies, end)) { ++ if (time_after(jiffies, end + (TIMEOUT_MS * HZ) / 1000)) { ++ return -1; ++ } ++ ++ if (need_resched()) ++ schedule(); ++ } ++#else ++ udelay(1); ++#endif ++ ++ return 0; ++} ++ ++#if !defined(CONFIG_PLATFORM_LS1012A) ++void pfe_ctrl_suspend(struct pfe_ctrl *ctrl) ++{ ++ int id; ++ ++ kthread_stop(ctrl->timer_thread); ++ ++ mutex_lock(&ctrl->mutex); ++ ++ initialized = 0; ++ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) ++ pe_dmem_write(id, cpu_to_be32(0x1), (unsigned long)virt_to_class_dmem(&class_resume), 4); ++ ++ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) { ++#if defined(CONFIG_PLATFORM_LS1012A) ++ if(id == TMU2_ID) continue; ++#endif ++ pe_dmem_write(id, cpu_to_be32(0x1), (unsigned long)virt_to_class_dmem(&tmu_resume), 4); ++ } ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ pe_dmem_write(UTIL_ID, cpu_to_be32(0x1), (unsigned long)virt_to_class_dmem(&util_resume), 4); ++#endif ++ ++ pe_sync_stop(&pfe->ctrl, 0xFF); ++ ++ mutex_unlock(&ctrl->mutex); ++} ++ ++void pfe_ctrl_resume(struct pfe_ctrl *ctrl) ++{ ++ mutex_lock(&ctrl->mutex); ++ initialized = 1; ++ pe_start(&pfe->ctrl, 0xFF); ++ mutex_unlock(&ctrl->mutex); ++ ++ ctrl->timer_thread = kthread_create(pfe_ctrl_timer, ctrl, "pfe_ctrl_timer"); ++ ++ wake_up_process(ctrl->timer_thread); ++} ++#endif ++ ++/** PE sync stop. ++* Stops packet processing for a list of PE's (specified using a bitmask). ++* The caller must hold ctrl->mutex. ++* ++* @param ctrl Control context ++* @param pe_mask Mask of PE id's to stop ++* ++*/ ++int pe_sync_stop(struct pfe_ctrl *ctrl, int pe_mask) ++{ ++ struct pe_sync_mailbox *mbox; ++ int pe_stopped = 0; ++ unsigned long end = jiffies + 2; ++ int i; ++ ++#if defined(CONFIG_PLATFORM_LS1012A) ++ //TODO Util should be removed after IPSec is ported ++ pe_mask &= 0x2FF; //Exclude Util + TMU2 ++#endif ++ for (i = 0; i < MAX_PE; i++) ++ if (pe_mask & (1 << i)) { ++ mbox = (void *)ctrl->sync_mailbox_baseaddr[i]; ++ ++ pe_dmem_write(i, cpu_to_be32(0x1), (unsigned long)&mbox->stop, 4); ++ } ++ ++ while (pe_stopped != pe_mask) { ++ for (i = 0; i < MAX_PE; i++) ++ if ((pe_mask & (1 << i)) && !(pe_stopped & (1 << i))) { ++ mbox = (void *)ctrl->sync_mailbox_baseaddr[i]; ++ ++ if (pe_dmem_read(i, (unsigned long)&mbox->stopped, 4) & cpu_to_be32(0x1)) ++ pe_stopped |= (1 << i); ++ } ++ ++ if (relax(end) < 0) ++ goto err; ++ } ++ ++ return 0; ++ ++err: ++ printk(KERN_ERR "%s: timeout, %x %x\n", __func__, pe_mask, pe_stopped); ++ ++ for (i = 0; i < MAX_PE; i++) ++ if (pe_mask & (1 << i)) { ++ mbox = (void *)ctrl->sync_mailbox_baseaddr[i]; ++ ++ pe_dmem_write(i, cpu_to_be32(0x0), (unsigned long)&mbox->stop, 4); ++ } ++ ++ return -EIO; ++} ++ ++/** PE start. ++* Starts packet processing for a list of PE's (specified using a bitmask). ++* The caller must hold ctrl->mutex. ++* ++* @param ctrl Control context ++* @param pe_mask Mask of PE id's to start ++* ++*/ ++void pe_start(struct pfe_ctrl *ctrl, int pe_mask) ++{ ++ struct pe_sync_mailbox *mbox; ++ int i; ++ ++#if defined(CONFIG_PLATFORM_LS1012A) ++ //TODO Util should be removed after IPSec is ported ++ pe_mask &= 0x2FF; //Exclude Util + TMU2 ++#endif ++ for (i = 0; i < MAX_PE; i++) ++ if (pe_mask & (1 << i)) { ++ ++ mbox = (void *)ctrl->sync_mailbox_baseaddr[i]; ++ ++ pe_dmem_write(i, cpu_to_be32(0x0), (unsigned long)&mbox->stop, 4); ++ } ++} ++ ++ ++/** Sends a control request to a given PE (to copy data to/from internal memory from/to DDR). ++* The caller must hold ctrl->mutex. ++* ++* @param ctrl Control context ++* @param id PE id ++* @param dst Physical destination address of data ++* @param src Physical source address of data ++* @param len Data length ++* ++*/ ++int pe_request(struct pfe_ctrl *ctrl, int id, unsigned short cmd_type, unsigned long dst, unsigned long src, int len) ++{ ++ struct pe_msg_mailbox mbox = { ++ .dst = cpu_to_be32(dst), ++ .src = cpu_to_be32(src), ++ .len = cpu_to_be32(len), ++ .request = cpu_to_be32((cmd_type << 16) | 0x1), ++ }; ++ struct pe_msg_mailbox *pmbox = (void *)ctrl->msg_mailbox_baseaddr[id]; ++ unsigned long end = jiffies + 2; ++ u32 rc; ++ ++ /* This works because .request is written last */ ++ pe_dmem_memcpy_to32(id, (unsigned long)pmbox, &mbox, sizeof(mbox)); ++ ++ while ((rc = pe_dmem_read(id, (unsigned long)&pmbox->request, 4)) & cpu_to_be32(0xffff)) { ++ if (relax(end) < 0) ++ goto err; ++ } ++ ++ rc = be32_to_cpu(rc); ++ ++ return rc >> 16; ++ ++err: ++ printk(KERN_ERR "%s: timeout, %x\n", __func__, be32_to_cpu(rc)); ++ pe_dmem_write(id, cpu_to_be32(0), (unsigned long)&pmbox->request, 4); ++ return -EIO; ++} ++ ++ ++/** Control code timer thread. ++* ++* A kernel thread is used so that the timer code can be run under the control path mutex. ++* The thread wakes up regularly and checks if any timer in the timer list as expired. ++* The timers are re-started automatically. ++* The code tries to keep the number of times a timer runs per unit time constant on average, ++* if the thread scheduling is delayed, it's possible for a particular timer to be scheduled in ++* quick succession to make up for the lost time. ++* ++* @param data Pointer to the control context structure ++* ++* @return 0 on sucess, a negative value on error ++* ++*/ ++static int pfe_ctrl_timer(void *data) ++{ ++ struct pfe_ctrl *ctrl = data; ++ TIMER_ENTRY *timer, *next; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ while (1) ++ { ++ schedule_timeout_uninterruptible(ctrl->timer_period); ++ ++ mutex_lock(&ctrl->mutex); ++ ++ list_for_each_entry_safe(timer, next, &ctrl->timer_list, list) ++ { ++ if (time_after(jiffies, timer->timeout)) ++ { ++ timer->timeout += timer->period; ++ ++ timer->handler(); ++ } ++ } ++ ++ mutex_unlock(&ctrl->mutex); ++ ++ if (kthread_should_stop()) ++ break; ++ } ++ ++ printk(KERN_INFO "%s exiting\n", __func__); ++ ++ return 0; ++} ++ ++ ++int pfe_ctrl_init(struct pfe *pfe) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ int id; ++ int rc; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ mutex_init(&ctrl->mutex); ++ spin_lock_init(&ctrl->lock); ++ ++ ctrl->timer_period = HZ / TIMER_TICKS_PER_SEC; ++ ++ INIT_LIST_HEAD(&ctrl->timer_list); ++ ++ /*INIT_WORK(&ctrl->work, comcerto_fpp_workqueue);*/ ++ ++ INIT_LIST_HEAD(&ctrl->msg_list); ++ ++ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) { ++ ctrl->sync_mailbox_baseaddr[id] = virt_to_class_dmem(&class_sync_mailbox); ++ ctrl->msg_mailbox_baseaddr[id] = virt_to_class_dmem(&class_msg_mailbox); ++ } ++ ++ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) { ++#if defined(CONFIG_PLATFORM_LS1012A) ++ if(id == TMU2_ID) continue; ++#endif ++ ctrl->sync_mailbox_baseaddr[id] = virt_to_tmu_dmem(&tmu_sync_mailbox); ++ ctrl->msg_mailbox_baseaddr[id] = virt_to_tmu_dmem(&tmu_msg_mailbox); ++ } ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ ctrl->sync_mailbox_baseaddr[UTIL_ID] = virt_to_util_dmem(&util_sync_mailbox); ++ ctrl->msg_mailbox_baseaddr[UTIL_ID] = virt_to_util_dmem(&util_msg_mailbox); ++#endif ++ ++ ctrl->hash_array_baseaddr = pfe->ddr_baseaddr + ROUTE_TABLE_BASEADDR; ++ ctrl->hash_array_phys_baseaddr = pfe->ddr_phys_baseaddr + ROUTE_TABLE_BASEADDR; ++ ctrl->ipsec_lmem_phys_baseaddr = CBUS_VIRT_TO_PFE(LMEM_BASE_ADDR + IPSEC_LMEM_BASEADDR); ++ ctrl->ipsec_lmem_baseaddr = (LMEM_BASE_ADDR + IPSEC_LMEM_BASEADDR); ++ ++ ctrl->timer_thread = kthread_create(pfe_ctrl_timer, ctrl, "pfe_ctrl_timer"); ++ if (IS_ERR(ctrl->timer_thread)) ++ { ++ printk (KERN_ERR "%s: kthread_create() failed\n", __func__); ++ rc = PTR_ERR(ctrl->timer_thread); ++ goto err0; ++ } ++ ++ ctrl->dev = pfe->dev; ++ ++ wake_up_process(ctrl->timer_thread); ++ ++ printk(KERN_INFO "%s finished\n", __func__); ++ ++ initialized = 1; ++ ++ return 0; ++ ++err0: ++ return rc; ++} ++ ++ ++void pfe_ctrl_exit(struct pfe *pfe) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ initialized = 0; ++ ++ kthread_stop(ctrl->timer_thread); ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_ctrl.h +@@ -0,0 +1,111 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_CTRL_H_ ++#define _PFE_CTRL_H_ ++ ++#include <linux/dmapool.h> ++ ++#include "pfe_mod.h" ++#include "pfe/pfe.h" ++ ++#define DMA_BUF_SIZE_128 0x80 /* enough for 1 conntracks */ ++#define DMA_BUF_SIZE_256 0x100 /* enough for 2 conntracks, 1 bridge entry or 1 multicast entry */ ++#define DMA_BUF_SIZE_512 0x200 /* 512bytes dma allocated buffers used by rtp relay feature */ ++#define DMA_BUF_MIN_ALIGNMENT 8 ++#define DMA_BUF_BOUNDARY (4 * 1024) /* bursts can not cross 4k boundary */ ++ ++#define CMD_TX_ENABLE 0x0501 ++#define CMD_TX_DISABLE 0x0502 ++ ++#define CMD_RX_LRO 0x0011 ++#define CMD_PKTCAP_ENABLE 0x0d01 ++#define CMD_QM_EXPT_RATE 0x020c ++ ++#define EXPT_TYPE_PCAP 0x3 ++ ++struct pfe_ctrl { ++ struct mutex mutex; ++ spinlock_t lock; ++ ++ void *dma_pool; ++ void *dma_pool_512; ++ void *dma_pool_128; ++ ++ struct device *dev; ++ ++ void *hash_array_baseaddr; /** Virtual base address of the conntrack hash array */ ++ unsigned long hash_array_phys_baseaddr; /** Physical base address of the conntrack hash array */ ++ ++ struct task_struct *timer_thread; ++ struct list_head timer_list; ++ unsigned long timer_period; ++ ++ int (*event_cb)(u16, u16, u16*); ++ ++ unsigned long sync_mailbox_baseaddr[MAX_PE]; /* Sync mailbox PFE internal address, initialized when parsing elf images */ ++ unsigned long msg_mailbox_baseaddr[MAX_PE]; /* Msg mailbox PFE internal address, initialized when parsing elf images */ ++ ++ unsigned long class_dmem_sh; ++ unsigned long class_pe_lmem_sh; ++ unsigned long tmu_dmem_sh; ++ unsigned long util_dmem_sh; ++ unsigned long util_ddr_sh; ++ struct clk *clk_axi; ++ unsigned int sys_clk; // AXI clock value, in KHz ++ void *ipsec_lmem_baseaddr; ++ unsigned long ipsec_lmem_phys_baseaddr; ++ ++ /* used for asynchronous message transfer to PFE */ ++ struct list_head msg_list; ++ struct work_struct work; ++}; ++ ++int pfe_ctrl_init(struct pfe *pfe); ++void pfe_ctrl_exit(struct pfe *pfe); ++ ++int pe_send_cmd(unsigned short cmd_type, unsigned short action, unsigned long param1, unsigned long param2); ++int pe_sync_stop(struct pfe_ctrl *ctrl, int pe_mask); ++void pe_start(struct pfe_ctrl *ctrl, int pe_mask); ++int pe_request(struct pfe_ctrl *ctrl, int id,unsigned short cmd_type, unsigned long dst, unsigned long src, int len); ++int pe_read(struct pfe_ctrl *ctrl, int id, u32 *dst, unsigned long src, int len, int clear_flag); ++int tmu_pe_request(struct pfe_ctrl *ctrl, int id, unsigned int tmu_cmd_bitmask); ++ ++int pfe_ctrl_set_eth_state(int id, unsigned int state, unsigned char *mac_addr); ++int pfe_ctrl_set_lro(char enable); ++#ifdef CFG_PCAP ++int pfe_ctrl_set_pcap(char enable); ++int pfe_ctrl_set_pcap_ratelimit(u32 pkts_per_msec); ++#endif ++void pfe_ctrl_suspend(struct pfe_ctrl *ctrl); ++void pfe_ctrl_resume(struct pfe_ctrl *ctrl); ++int relax(unsigned long end); ++ ++/* used for asynchronous message transfer to PFE */ ++#define FPP_MAX_MSG_LENGTH 256 /* expressed in U8 -> 256 bytes*/ ++struct fpp_msg { ++ struct list_head list; ++ void (*callback)(unsigned long, int, u16, u16 *); ++ unsigned long data; ++ u16 fcode; ++ u16 length; ++ u16 *payload; ++}; ++ ++#endif /* _PFE_CTRL_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_ctrl_hal.c +@@ -0,0 +1,207 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++/* OS abstraction functions used by PFE control code */ ++ ++#include <linux/slab.h> ++ ++#include "pfe_ctrl_hal.h" ++ ++#include "pfe_mod.h" ++ ++extern char *__class_dmem_sh; ++extern char *__tmu_dmem_sh; ++#if !defined(CONFIG_UTIL_DISABLED) ++extern char *__util_dmem_sh; ++extern char *__util_ddr_sh; ++#endif ++ ++HostMessage msg_buf; ++static int msg_buf_used = 0; ++unsigned long virt_to_class_dmem(void *p) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ if (p) ++ return (unsigned long)p - (unsigned long)&__class_dmem_sh + ctrl->class_dmem_sh; ++ else ++ return 0; ++} ++unsigned long virt_to_tmu_dmem(void *p) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ if (p) ++ return (unsigned long)p - (unsigned long)&__tmu_dmem_sh + ctrl->tmu_dmem_sh; ++ else ++ return 0; ++} ++ ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++unsigned long virt_to_util_dmem(void *p) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ if (p) ++ return (unsigned long)p - (unsigned long)&__util_dmem_sh + ctrl->util_dmem_sh; ++ else ++ return 0; ++} ++ ++/** Returns the DDR physical address of a Util PE shared DDR variable. ++ * ++ * @param p pointer (kernel space, virtual) to be converted to a physical address. ++ */ ++unsigned long virt_to_util_ddr(void *p) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ if (p) ++ return (unsigned long)p - (unsigned long)&__util_ddr_sh + ctrl->util_ddr_sh; ++ else ++ return 0; ++} ++/** Returns the virtual address of a Util PE shared DDR variable. ++ * ++ * @param p pointer (kernel space, virtual) to be converted to a pointer (usable in kernel space) ++ * pointing to the actual data. ++ */ ++ ++void * virt_to_util_virt(void *p) ++{ ++ if (p) ++ return DDR_PHYS_TO_VIRT(virt_to_util_ddr(p)); ++ else ++ return NULL; ++} ++#endif ++unsigned long virt_to_phys_iram(void *p) ++{ ++ if (p) ++ return (p - pfe->iram_baseaddr) + pfe->iram_phys_baseaddr; ++ else ++ return 0; ++} ++ ++unsigned long virt_to_phys_ipsec_lmem(void *p) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ if (p) ++ return (p - ctrl->ipsec_lmem_baseaddr) + ctrl->ipsec_lmem_phys_baseaddr; ++ else ++ return 0; ++} ++ ++unsigned long virt_to_phys_ipsec_axi(void *p) ++{ ++ if (p) ++ return (p - pfe->ipsec_baseaddr) + pfe->ipsec_phys_baseaddr; ++ else ++ return 0; ++} ++ ++ ++HostMessage *msg_alloc(void) ++{ ++ if (msg_buf_used) ++ { ++ printk(KERN_ERR "%s: failed\n", __func__); ++ return NULL; ++ } ++ ++ msg_buf_used = 1; ++ ++ return &msg_buf; ++} ++ ++void msg_free(HostMessage *msg) ++{ ++ if (!msg_buf_used) ++ printk(KERN_ERR "%s: freing already free msg buffer\n", __func__); ++ ++ msg_buf_used = 0; ++} ++ ++int msg_send(HostMessage *msg) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ int rc = -1; ++ ++ if (!ctrl->event_cb) ++ goto out; ++ ++ if (ctrl->event_cb(msg->code, msg->length, msg->data) < 0) ++ goto out; ++ ++ rc = 0; ++ ++out: ++ msg_free(msg); ++ ++ return rc; ++} ++ ++ ++void timer_init(TIMER_ENTRY *timer, TIMER_HANDLER handler) ++{ ++ timer->handler = handler; ++ timer->running = 0; ++} ++ ++ ++void timer_add(TIMER_ENTRY *timer, u16 granularity) ++{ ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ ++ timer->period = granularity; ++ timer->timeout = jiffies + timer->period; ++ ++ if (!timer->running) ++ { ++ list_add(&timer->list, &ctrl->timer_list); ++ timer->running = 1; ++ } ++} ++ ++ ++void timer_del(TIMER_ENTRY *timer) ++{ ++ ++ if (timer->running) ++ { ++ list_del(&timer->list); ++ timer->running = 0; ++ } ++} ++ ++ ++void *Heap_Alloc(int size) ++{ ++ /* FIXME we may want to use dma API's and use non cacheable memory */ ++ return pfe_kmalloc(size, GFP_KERNEL); ++} ++ ++ ++void Heap_Free(void *p) ++{ ++ pfe_kfree(p); ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_ctrl_hal.h +@@ -0,0 +1,129 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_CTRL_HAL_H_ ++#define _PFE_CTRL_HAL_H_ ++ ++#include <linux/timer.h> ++#include <linux/jiffies.h> ++#include <linux/string.h> ++#include <linux/elf.h> ++#include <linux/slab.h> ++#include <asm/byteorder.h> ++#include <asm/io.h> ++ ++#include "pfe_mod.h" ++ ++#define CLASS_DMEM_SH(var) __attribute__((section(".class_dmem_sh_" #var))) var ++#define CLASS_PE_LMEM_SH(var) __attribute__((section(".class_pe_lmem_sh_" #var))) var ++#define TMU_DMEM_SH(var) __attribute__((section(".tmu_dmem_sh_" #var))) var ++#define UTIL_DMEM_SH(var) __attribute__((section(".util_dmem_sh_" #var))) var ++#define UTIL_DDR_SH(var) __attribute__((section(".util_ddr_sh_" #var))) var ++ ++#define CLASS_DMEM_SH2(var) __attribute__((section(".class_dmem_sh_" #var))) class_##var ++#define CLASS_PE_LMEM_SH2(var) __attribute__((section(".class_pe_lmem_sh_" #var))) class_##var ++#define TMU_DMEM_SH2(var) __attribute__((section(".tmu_dmem_sh_" #var))) tmu_##var ++#define UTIL_DMEM_SH2(var) __attribute__((section(".util_dmem_sh_" #var))) util_##var ++ ++/** Translate the name of a shared variable to its PFE counterpart. ++ * Those macros may be used to determine the address of a shared variable, ++ * and will work even if the variable is accessed through a macro, as is the case ++ * with most fields of gFppGlobals. ++ */ ++#define CONCAT(str, var) str##var ++#define CLASS_VARNAME2(var) CONCAT(class_, var) ++#define UTIL_VARNAME2(var) CONCAT(util_, var) ++#define TMU_VARNAME2(var) CONCAT(tmu_, var) ++ ++typedef struct tHostMessage { ++ u16 length; ++ u16 code; ++ u16 data[128]; ++} HostMessage; ++ ++HostMessage *msg_alloc(void); ++void msg_free(HostMessage *msg); ++int msg_send(HostMessage *msg); ++ ++ ++unsigned long virt_to_class(void *p); ++unsigned long virt_to_class_dmem(void *p); ++unsigned long virt_to_class_pe_lmem(void *p); ++unsigned long virt_to_tmu_dmem(void *p); ++unsigned long virt_to_util_dmem(void *p); ++unsigned long virt_to_util_ddr(void *p); ++void * virt_to_util_virt(void *p); ++unsigned long virt_to_phys_iram(void *p); ++unsigned long virt_to_phys_ipsec_lmem(void *p); ++unsigned long virt_to_phys_ipsec_axi(void *p); ++ ++ ++#define TIMER_TICKS_PER_SEC 100 ++ ++#if TIMER_TICKS_PER_SEC > HZ ++#error TIMER_TICKS_PER_SEC is too high ++#endif ++ ++ ++typedef void (* TIMER_HANDLER)(void); ++ ++typedef struct { ++ struct list_head list; ++ unsigned long timeout; ++ unsigned long period; ++ TIMER_HANDLER handler; ++ char running; ++} TIMER_ENTRY; ++ ++ ++/** Initializes a timer structure. ++* Must be called once for each TIMER_ENTRY structure. ++* The caller must be holding the ctrl->mutex. ++* ++* @param timer pointer to the timer to be initialized ++* @param handler timer handler function pointer ++* ++*/ ++void timer_init(TIMER_ENTRY *timer, TIMER_HANDLER handler); ++ ++/** Adds a timer to the running timer list. ++* It's safe to call even if the timer was already running. In this case we just update the granularity. ++* The caller must be holding the ctrl->mutex. ++* ++* @param timer pointer to the timer to be added ++* @param granularity granularity of the timer (in timer tick units) ++* ++*/ ++void timer_add(TIMER_ENTRY *timer, u16 granularity); ++ ++/** Deletes a timer from the running timer list. ++* It's safe to call even if the timer is no longer running. ++* The caller must be holding the ctrl->mutex. ++* ++* @param timer pointer to the timer to be removed ++*/ ++void timer_del(TIMER_ENTRY *timer); ++ ++void *Heap_Alloc(int size); ++ ++#define Heap_Alloc_ARAM(s) Heap_Alloc(s) ++#define __Heap_Alloc(h, s) Heap_Alloc(s) ++void Heap_Free(void *p); ++ ++#endif /* _PFE_CTRL_HAL_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_debugfs.c +@@ -0,0 +1,109 @@ ++/* ++ * (C) Copyright 2013 ++ * Author : Freescale Technologes ++ * ++ * See file CREDITS for list of people who contributed to this ++ * project. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License as ++ * published by the Free Software Foundation; either version 2 of ++ * the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, ++ * MA 02111-1307 USA ++ * */ ++ ++#include <linux/module.h> ++#include <linux/debugfs.h> ++#include <linux/platform_device.h> ++ ++#include "pfe_mod.h" ++ ++static int dmem_show(struct seq_file *s, void *unused) ++{ ++ u32 dmem_addr, val; ++ int id = (long int)s->private; ++ int i; ++ ++ for (dmem_addr = 0; dmem_addr < CLASS_DMEM_SIZE; dmem_addr += 8 * 4) { ++ seq_printf(s, "%04x:", dmem_addr); ++ ++ for (i = 0; i < 8; i++) { ++ val = pe_dmem_read(id, dmem_addr + i * 4, 4); ++ seq_printf(s, " %02x %02x %02x %02x", val & 0xff, (val >> 8) & 0xff, (val >> 16) & 0xff, (val >> 24) & 0xff); ++ } ++ ++ seq_printf(s, "\n"); ++ } ++ ++ return 0; ++} ++ ++static int dmem_open(struct inode *inode, struct file *file) ++{ ++ return single_open(file, dmem_show, inode->i_private); ++} ++ ++static const struct file_operations dmem_fops = { ++ .open = dmem_open, ++ .read = seq_read, ++ .llseek = seq_lseek, ++ .release = single_release, ++}; ++ ++int pfe_debugfs_init(struct pfe *pfe) ++{ ++ struct dentry *d; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ pfe->dentry = debugfs_create_dir("pfe", NULL); ++ if (IS_ERR_OR_NULL(pfe->dentry)) ++ goto err_dir; ++ ++ d = debugfs_create_file("pe0_dmem", S_IRUGO, pfe->dentry, (void *)0, &dmem_fops); ++ if (IS_ERR_OR_NULL(d)) ++ goto err_pe; ++ ++ d = debugfs_create_file("pe1_dmem", S_IRUGO, pfe->dentry, (void *)1, &dmem_fops); ++ if (IS_ERR_OR_NULL(d)) ++ goto err_pe; ++ ++ d = debugfs_create_file("pe2_dmem", S_IRUGO, pfe->dentry, (void *)2, &dmem_fops); ++ if (IS_ERR_OR_NULL(d)) ++ goto err_pe; ++ ++ d = debugfs_create_file("pe3_dmem", S_IRUGO, pfe->dentry, (void *)3, &dmem_fops); ++ if (IS_ERR_OR_NULL(d)) ++ goto err_pe; ++ ++ d = debugfs_create_file("pe4_dmem", S_IRUGO, pfe->dentry, (void *)4, &dmem_fops); ++ if (IS_ERR_OR_NULL(d)) ++ goto err_pe; ++ ++ d = debugfs_create_file("pe5_dmem", S_IRUGO, pfe->dentry, (void *)5, &dmem_fops); ++ if (IS_ERR_OR_NULL(d)) ++ goto err_pe; ++ ++ return 0; ++ ++err_pe: ++ debugfs_remove_recursive(pfe->dentry); ++ ++err_dir: ++ return -1; ++} ++ ++void pfe_debugfs_exit(struct pfe *pfe) ++{ ++ debugfs_remove_recursive(pfe->dentry); ++} ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_debugfs.h +@@ -0,0 +1,8 @@ ++#ifndef _PFE_DEBUGFS_H_ ++#define _PFE_DEBUGFS_H_ ++ ++int pfe_debugfs_init(struct pfe *pfe); ++void pfe_debugfs_exit(struct pfe *pfe); ++#endif /* _PFE_DEBUGFS_H_ */ ++ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_eth.c +@@ -0,0 +1,2956 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++/** @pfe_eth.c. ++ * Ethernet driver for to handle exception path for PFE. ++ * - uses HIF functions to send/receive packets. ++ * - uses ctrl function to start/stop interfaces. ++ * - uses direct register accesses to control phy operation. ++ */ ++#include <linux/version.h> ++#include <linux/kernel.h> ++#include <linux/interrupt.h> ++#include <linux/dma-mapping.h> ++#include <linux/dmapool.h> ++#include <linux/netdevice.h> ++#include <linux/etherdevice.h> ++#include <linux/ethtool.h> ++#include <linux/mii.h> ++#include <linux/phy.h> ++#include <linux/timer.h> ++#include <linux/hrtimer.h> ++#include <linux/platform_device.h> ++ ++#include <net/ip.h> ++#include <net/sock.h> ++ ++#include <asm/io.h> ++#include <asm/irq.h> ++#include <asm/delay.h> ++#include <linux/regmap.h> ++#include <linux/i2c.h> ++ ++#if defined(CONFIG_NF_CONNTRACK_MARK) ++#include <net/netfilter/nf_conntrack.h> ++#endif ++ ++#include "pfe_mod.h" ++#include "pfe_eth.h" ++ ++const char comcerto_eth_driver_version[]="1.0"; ++static void *cbus_emac_base[3]; ++static void *cbus_gpi_base[3]; ++ ++/* Forward Declaration */ ++static void pfe_eth_exit_one(struct pfe_eth_priv_s *priv); ++static void pfe_eth_flush_tx(struct pfe_eth_priv_s *priv, int force); ++static void pfe_eth_flush_txQ(struct pfe_eth_priv_s *priv, int txQ_num, int from_tx, int n_desc); ++ ++#if defined(CONFIG_PLATFORM_C2000) ++static void pfe_eth_set_device_wakeup(struct pfe *pfe); ++ ++ ++ ++unsigned int gemac_regs[] = { ++ 0x0000, /* Network control */ ++ 0x0004, /* Network configuration */ ++ 0x0008, /* Network status */ ++ 0x0010, /* DMA configuration */ ++ 0x0014, /* Transmit status */ ++ 0x0020, /* Receive status */ ++ 0x0024, /* Interrupt status */ ++ 0x0030, /* Interrupt mask */ ++ 0x0038, /* Received pause quantum */ ++ 0x003c, /* Transmit pause quantum */ ++ 0x0080, /* Hash register bottom [31:0] */ ++ 0x0084, /* Hash register bottom [63:32] */ ++ 0x0088, /* Specific address 1 bottom [31:0] */ ++ 0x008c, /* Specific address 1 top [47:32] */ ++ 0x0090, /* Specific address 2 bottom [31:0] */ ++ 0x0094, /* Specific address 2 top [47:32] */ ++ 0x0098, /* Specific address 3 bottom [31:0] */ ++ 0x009c, /* Specific address 3 top [47:32] */ ++ 0x00a0, /* Specific address 4 bottom [31:0] */ ++ 0x00a4, /* Specific address 4 top [47:32] */ ++ 0x00a8, /* Type ID Match 1 */ ++ 0x00ac, /* Type ID Match 2 */ ++ 0x00b0, /* Type ID Match 3 */ ++ 0x00b4, /* Type ID Match 4 */ ++ 0x00b8, /* Wake Up ON LAN */ ++ 0x00bc, /* IPG stretch register */ ++ 0x00c0, /* Stacked VLAN Register */ ++ 0x00fc, /* Module ID */ ++ 0x07a0 /* EMAC Control register */ ++}; ++#else ++unsigned int gemac_regs[] = { ++ 0x0004, /*Interrupt event */ ++ 0x0008, /*Interrupt mask */ ++ 0x0024, /*Ethernet control */ ++ 0x0064, /*MIB Control/Status */ ++ 0x0084, /*Receive control/status */ ++ 0x00C4, /*Transmit control */ ++ 0x00E4, /*Physical address low */ ++ 0x00E8, /*Physical address high */ ++ 0x0144, /*Transmit FIFO Watermark and Store and Forward Control*/ ++ 0x0190, /* Receive FIFO Section Full Threshold */ ++ 0x01A0, /* Transmit FIFO Section Empty Threshold */ ++ 0x01B0, /* Frame Truncation Length */ ++}; ++#endif ++/********************************************************************/ ++/* SYSFS INTERFACE */ ++/********************************************************************/ ++ ++ ++ ++#ifdef PFE_ETH_NAPI_STATS ++/* ++ * pfe_eth_show_napi_stats ++ */ ++static ssize_t pfe_eth_show_napi_stats(struct device *dev, ++ struct device_attribute *attr, char *buf) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ ssize_t len = 0; ++ ++ len += sprintf(buf + len, "sched: %u\n", priv->napi_counters[NAPI_SCHED_COUNT]); ++ len += sprintf(buf + len, "poll: %u\n", priv->napi_counters[NAPI_POLL_COUNT]); ++ len += sprintf(buf + len, "packet: %u\n", priv->napi_counters[NAPI_PACKET_COUNT]); ++ len += sprintf(buf + len, "budget: %u\n", priv->napi_counters[NAPI_FULL_BUDGET_COUNT]); ++ len += sprintf(buf + len, "desc: %u\n", priv->napi_counters[NAPI_DESC_COUNT]); ++ ++ return len; ++} ++ ++/* ++ * pfe_eth_set_napi_stats ++ */ ++static ssize_t pfe_eth_set_napi_stats(struct device *dev, ++ struct device_attribute *attr, const char *buf, size_t count) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ ++ memset(priv->napi_counters, 0, sizeof(priv->napi_counters)); ++ ++ return count; ++} ++#endif ++#ifdef PFE_ETH_TX_STATS ++/** pfe_eth_show_tx_stats ++ * ++ */ ++static ssize_t pfe_eth_show_tx_stats(struct device *dev, ++ struct device_attribute *attr, char *buf) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ ssize_t len = 0; ++ int i; ++ ++ len += sprintf(buf + len, "TX queues stats:\n"); ++ ++ for (i = 0; i < emac_txq_cnt; i++) { ++ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->dev, i); ++ ++ len += sprintf(buf + len, "\n"); ++ __netif_tx_lock_bh(tx_queue); ++ ++ hif_tx_lock(&pfe->hif); ++ len += sprintf(buf + len, "Queue %2d : credits = %10d\n", i, hif_lib_tx_credit_avail(pfe, priv->id, i)); ++ len += sprintf(buf + len, " tx packets = %10d\n", pfe->tmu_credit.tx_packets[priv->id][i]); ++ hif_tx_unlock(&pfe->hif); ++ ++ /* Don't output additionnal stats if queue never used */ ++ if (!pfe->tmu_credit.tx_packets[priv->id][i]) ++ goto skip; ++ ++ len += sprintf(buf + len, " clean_fail = %10d\n", priv->clean_fail[i]); ++ len += sprintf(buf + len, " stop_queue = %10d\n", priv->stop_queue_total[i]); ++ len += sprintf(buf + len, " stop_queue_hif = %10d\n", priv->stop_queue_hif[i]); ++ len += sprintf(buf + len, " stop_queue_hif_client = %10d\n", priv->stop_queue_hif_client[i]); ++ len += sprintf(buf + len, " stop_queue_credit = %10d\n", priv->stop_queue_credit[i]); ++skip: ++ __netif_tx_unlock_bh(tx_queue); ++ } ++ return len; ++} ++ ++/** pfe_eth_set_tx_stats ++ * ++ */ ++static ssize_t pfe_eth_set_tx_stats(struct device *dev, ++ struct device_attribute *attr, const char *buf, size_t count) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ int i; ++ ++ for (i = 0; i < emac_txq_cnt; i++) { ++ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->dev, i); ++ ++ __netif_tx_lock_bh(tx_queue); ++ priv->clean_fail[i] = 0; ++ priv->stop_queue_total[i] = 0; ++ priv->stop_queue_hif[i] = 0; ++ priv->stop_queue_hif_client[i]= 0; ++ priv->stop_queue_credit[i] = 0; ++ __netif_tx_unlock_bh(tx_queue); ++ } ++ ++ return count; ++} ++#endif ++/** pfe_eth_show_txavail ++ * ++ */ ++static ssize_t pfe_eth_show_txavail(struct device *dev, ++ struct device_attribute *attr, char *buf) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ ssize_t len = 0; ++ int i; ++ ++ for (i = 0; i < emac_txq_cnt; i++) { ++ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->dev, i); ++ ++ __netif_tx_lock_bh(tx_queue); ++ ++ len += sprintf(buf + len, "%d", hif_lib_tx_avail(&priv->client, i)); ++ ++ __netif_tx_unlock_bh(tx_queue); ++ ++ if (i == (emac_txq_cnt - 1)) ++ len += sprintf(buf + len, "\n"); ++ else ++ len += sprintf(buf + len, " "); ++ } ++ ++ return len; ++} ++ ++ ++/** pfe_eth_show_default_priority ++ * ++ */ ++static ssize_t pfe_eth_show_default_priority(struct device *dev, ++ struct device_attribute *attr, ++ char *buf) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ unsigned long flags; ++ int rc; ++ ++ spin_lock_irqsave(&priv->lock, flags); ++ rc = sprintf(buf, "%d\n", priv->default_priority); ++ spin_unlock_irqrestore(&priv->lock, flags); ++ ++ return rc; ++} ++ ++/** pfe_eth_set_default_priority ++ * ++ */ ++ ++static ssize_t pfe_eth_set_default_priority(struct device *dev, ++ struct device_attribute *attr, ++ const char *buf, size_t count) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(to_net_dev(dev)); ++ unsigned long flags; ++ ++ spin_lock_irqsave(&priv->lock, flags); ++ priv->default_priority = simple_strtoul(buf, NULL, 0); ++ spin_unlock_irqrestore(&priv->lock, flags); ++ ++ return count; ++} ++ ++static DEVICE_ATTR(txavail, 0444, pfe_eth_show_txavail, NULL); ++static DEVICE_ATTR(default_priority, 0644, pfe_eth_show_default_priority, pfe_eth_set_default_priority); ++ ++#ifdef PFE_ETH_NAPI_STATS ++static DEVICE_ATTR(napi_stats, 0644, pfe_eth_show_napi_stats, pfe_eth_set_napi_stats); ++#endif ++ ++#ifdef PFE_ETH_TX_STATS ++static DEVICE_ATTR(tx_stats, 0644, pfe_eth_show_tx_stats, pfe_eth_set_tx_stats); ++#endif ++ ++ ++/** pfe_eth_sysfs_init ++ * ++ */ ++static int pfe_eth_sysfs_init(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int err; ++ ++ /* Initialize the default values */ ++ /* By default, packets without conntrack will use this default high priority queue */ ++ priv->default_priority = 15; ++ ++ /* Create our sysfs files */ ++ err = device_create_file(&dev->dev, &dev_attr_default_priority); ++ if (err) { ++ netdev_err(dev, "failed to create default_priority sysfs files\n"); ++ goto err_priority; ++ } ++ ++ err = device_create_file(&dev->dev, &dev_attr_txavail); ++ if (err) { ++ netdev_err(dev, "failed to create default_priority sysfs files\n"); ++ goto err_txavail; ++ } ++ ++#ifdef PFE_ETH_NAPI_STATS ++ err = device_create_file(&dev->dev, &dev_attr_napi_stats); ++ if (err) { ++ netdev_err(dev, "failed to create napi stats sysfs files\n"); ++ goto err_napi; ++ } ++#endif ++ ++#ifdef PFE_ETH_TX_STATS ++ err = device_create_file(&dev->dev, &dev_attr_tx_stats); ++ if (err) { ++ netdev_err(dev, "failed to create tx stats sysfs files\n"); ++ goto err_tx; ++ } ++#endif ++ ++ return 0; ++ ++#ifdef PFE_ETH_TX_STATS ++err_tx: ++#endif ++#ifdef PFE_ETH_NAPI_STATS ++ device_remove_file(&dev->dev, &dev_attr_napi_stats); ++ ++err_napi: ++#endif ++ device_remove_file(&dev->dev, &dev_attr_txavail); ++ ++err_txavail: ++ device_remove_file(&dev->dev, &dev_attr_default_priority); ++ ++err_priority: ++ return -1; ++} ++ ++/** pfe_eth_sysfs_exit ++ * ++ */ ++void pfe_eth_sysfs_exit(struct net_device *dev) ++{ ++#ifdef PFE_ETH_TX_STATS ++ device_remove_file(&dev->dev, &dev_attr_tx_stats); ++#endif ++ ++#ifdef PFE_ETH_NAPI_STATS ++ device_remove_file(&dev->dev, &dev_attr_napi_stats); ++#endif ++ device_remove_file(&dev->dev, &dev_attr_txavail); ++ device_remove_file(&dev->dev, &dev_attr_default_priority); ++} ++ ++/*************************************************************************/ ++/* ETHTOOL INTERCAE */ ++/*************************************************************************/ ++ ++#if defined(CONFIG_PLATFORM_C2000) ++static char stat_gstrings[][ETH_GSTRING_LEN] = { ++ "tx- octets", ++ "tx- packets", ++ "tx- broadcast", ++ "tx- multicast", ++ "tx- pause", ++ "tx- 64 bytes packets", ++ "tx- 64 - 127 bytes packets", ++ "tx- 128 - 255 bytes packets", ++ "tx- 256 - 511 bytes packets", ++ "tx- 512 - 1023 bytes packets", ++ "tx- 1024 - 1518 bytes packets", ++ "tx- > 1518 bytes packets", ++ "tx- underruns - errors", ++ "tx- single collision", ++ "tx- multi collision", ++ "tx- exces. collision - errors", ++ "tx- late collision - errors", ++ "tx- deferred", ++ "tx- carrier sense - errors", ++ "rx- octets", ++ "rx- packets", ++ "rx- broadcast", ++ "rx- multicast", ++ "rx- pause", ++ "rx- 64 bytes packets", ++ "rx- 64 - 127 bytes packets", ++ "rx- 128 - 255 bytes packets", ++ "rx- 256 - 511 bytes packets", ++ "rx- 512 - 1023 bytes packets", ++ "rx- 1024 - 1518 bytes packets", ++ "rx- > 1518 bytes packets", ++ "rx- undersize -errors", ++ "rx- oversize - errors ", ++ "rx- jabbers - errors", ++ "rx- fcs - errors", ++ "rx- length - errors", ++ "rx- symbol - errors", ++ "rx- align - errors", ++ "rx- ressource - errors", ++ "rx- overrun - errors", ++ "rx- IP cksum - errors", ++ "rx- TCP cksum - errors", ++ "rx- UDP cksum - errors" ++}; ++ ++ ++/** ++ * pfe_eth_gstrings - Fill in a buffer with the strings which correspond to ++ * the stats. ++ * ++ */ ++static void pfe_eth_gstrings(struct net_device *dev, u32 stringset, u8 * buf) ++{ ++ switch (stringset) { ++ case ETH_SS_STATS: ++ memcpy(buf, stat_gstrings, (EMAC_RMON_LEN - 2) * ETH_GSTRING_LEN); ++ break; ++ ++ default: ++ WARN_ON(1); ++ break; ++ } ++} ++ ++/** ++ * pfe_eth_fill_stats - Fill in an array of 64-bit statistics from ++ * various sources. This array will be appended ++ * to the end of the ethtool_stats* structure, and ++ * returned to user space ++ */ ++static void pfe_eth_fill_stats(struct net_device *dev, struct ethtool_stats *dummy, u64 * buf) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int i; ++ for (i=0;i<EMAC_RMON_LEN;i++, buf++) { ++ *buf = readl(priv->EMAC_baseaddr + EMAC_RMON_BASE_OFST + (i << 2)); ++ if ( ( i == EMAC_RMON_TXBYTES_POS ) || ( i == EMAC_RMON_RXBYTES_POS ) ){ ++ i++; ++ *buf |= (u64)readl(priv->EMAC_baseaddr + EMAC_RMON_BASE_OFST + (i << 2)) << 32; ++ } ++ } ++ ++} ++ ++/** ++ * pfe_eth_stats_count - Returns the number of stats (and their corresponding strings) ++ * ++ */ ++static int pfe_eth_stats_count(struct net_device *dev, int sset) ++{ ++ switch (sset) { ++ case ETH_SS_STATS: ++ return EMAC_RMON_LEN - 2; ++ default: ++ return -EOPNOTSUPP; ++ } ++} ++ ++#if defined(CONFIG_PLATFORM_C2000) ++/** ++ * pfe_eth_set_wol - Set the magic packet option, in WoL register. ++ * ++ */ ++static int pfe_eth_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ if (wol->wolopts & ~(WAKE_MAGIC | WAKE_ARP | WAKE_MCAST | WAKE_UCAST)) ++ return -EOPNOTSUPP; ++ ++ priv->wol = 0; ++ ++ if (wol->wolopts & WAKE_MAGIC) ++ priv->wol |= EMAC_WOL_MAGIC; ++ if (wol->wolopts & WAKE_ARP) ++ priv->wol |= EMAC_WOL_ARP; ++ if (wol->wolopts & WAKE_MCAST) ++ priv->wol |= EMAC_WOL_MULTI; ++ if (wol->wolopts & WAKE_UCAST) ++ priv->wol |= EMAC_WOL_SPEC_ADDR; ++ ++ pfe_eth_set_device_wakeup(priv->pfe); ++ ++ return 0; ++} ++ ++/** ++ * ++ * pfe_eth_get_wol - Get the WoL options. ++ * ++ */ ++static void pfe_eth_get_wol(struct net_device *dev, struct ethtool_wolinfo *wol) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ wol->supported = (WAKE_MAGIC | WAKE_ARP | WAKE_MCAST | WAKE_UCAST); ++ wol->wolopts = 0; ++ ++ if(priv->wol & EMAC_WOL_MAGIC) ++ wol->wolopts |= WAKE_MAGIC; ++ if(priv->wol & EMAC_WOL_ARP) ++ wol->wolopts |= WAKE_ARP; ++ if(priv->wol & EMAC_WOL_MULTI) ++ wol->wolopts |= WAKE_UCAST; ++ if(priv->wol & EMAC_WOL_SPEC_ADDR) ++ wol->wolopts |= WAKE_UCAST; ++ ++ memset(&wol->sopass, 0, sizeof(wol->sopass)); ++} ++#endif ++/** ++ * pfe_eth_gemac_reglen - Return the length of the register structure. ++ * ++ */ ++static int pfe_eth_gemac_reglen(struct net_device *dev) ++{ ++ return (sizeof (gemac_regs)/ sizeof(u32)) + (( MAX_UC_SPEC_ADDR_REG - 3 ) * 2); ++} ++ ++/** ++ * pfe_eth_gemac_get_regs - Return the gemac register structure. ++ * ++ */ ++static void pfe_eth_gemac_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *regbuf) ++{ ++ int i,j; ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ u32 *buf = (u32 *) regbuf; ++ ++ for (i = 0; i < sizeof (gemac_regs) / sizeof (u32); i++) ++ buf[i] = readl( priv->EMAC_baseaddr + gemac_regs[i] ); ++ ++ for (j = 0; j < (( MAX_UC_SPEC_ADDR_REG - 3 ) * 2); j++,i++) ++ buf[i] = readl( priv->EMAC_baseaddr + EMAC_SPEC5_ADD_BOT + (j<<2) ); ++ ++} ++ ++ ++#else //if defined(CONFIG_PLATFORM_C2000) ++/*MTIP GEMAC */ ++static const struct fec_stat { ++ char name[ETH_GSTRING_LEN]; ++ u16 offset; ++} fec_stats[] = { ++ /* RMON TX */ ++ { "tx_dropped", RMON_T_DROP }, ++ { "tx_packets", RMON_T_PACKETS }, ++ { "tx_broadcast", RMON_T_BC_PKT }, ++ { "tx_multicast", RMON_T_MC_PKT }, ++ { "tx_crc_errors", RMON_T_CRC_ALIGN }, ++ { "tx_undersize", RMON_T_UNDERSIZE }, ++ { "tx_oversize", RMON_T_OVERSIZE }, ++ { "tx_fragment", RMON_T_FRAG }, ++ { "tx_jabber", RMON_T_JAB }, ++ { "tx_collision", RMON_T_COL }, ++ { "tx_64byte", RMON_T_P64 }, ++ { "tx_65to127byte", RMON_T_P65TO127 }, ++ { "tx_128to255byte", RMON_T_P128TO255 }, ++ { "tx_256to511byte", RMON_T_P256TO511 }, ++ { "tx_512to1023byte", RMON_T_P512TO1023 }, ++ { "tx_1024to2047byte", RMON_T_P1024TO2047 }, ++ { "tx_GTE2048byte", RMON_T_P_GTE2048 }, ++ { "tx_octets", RMON_T_OCTETS }, ++ ++ /* IEEE TX */ ++ { "IEEE_tx_drop", IEEE_T_DROP }, ++ { "IEEE_tx_frame_ok", IEEE_T_FRAME_OK }, ++ { "IEEE_tx_1col", IEEE_T_1COL }, ++ { "IEEE_tx_mcol", IEEE_T_MCOL }, ++ { "IEEE_tx_def", IEEE_T_DEF }, ++ { "IEEE_tx_lcol", IEEE_T_LCOL }, ++ { "IEEE_tx_excol", IEEE_T_EXCOL }, ++ { "IEEE_tx_macerr", IEEE_T_MACERR }, ++ { "IEEE_tx_cserr", IEEE_T_CSERR }, ++ { "IEEE_tx_sqe", IEEE_T_SQE }, ++ { "IEEE_tx_fdxfc", IEEE_T_FDXFC }, ++ { "IEEE_tx_octets_ok", IEEE_T_OCTETS_OK }, ++ ++ /* RMON RX */ ++ { "rx_packets", RMON_R_PACKETS }, ++ { "rx_broadcast", RMON_R_BC_PKT }, ++ { "rx_multicast", RMON_R_MC_PKT }, ++ { "rx_crc_errors", RMON_R_CRC_ALIGN }, ++ { "rx_undersize", RMON_R_UNDERSIZE }, ++ { "rx_oversize", RMON_R_OVERSIZE }, ++ { "rx_fragment", RMON_R_FRAG }, ++ { "rx_jabber", RMON_R_JAB }, ++ { "rx_64byte", RMON_R_P64 }, ++ { "rx_65to127byte", RMON_R_P65TO127 }, ++ { "rx_128to255byte", RMON_R_P128TO255 }, ++ { "rx_256to511byte", RMON_R_P256TO511 }, ++ { "rx_512to1023byte", RMON_R_P512TO1023 }, ++ { "rx_1024to2047byte", RMON_R_P1024TO2047 }, ++ { "rx_GTE2048byte", RMON_R_P_GTE2048 }, ++ { "rx_octets", RMON_R_OCTETS }, ++ ++ /* IEEE RX */ ++ { "IEEE_rx_drop", IEEE_R_DROP }, ++ { "IEEE_rx_frame_ok", IEEE_R_FRAME_OK }, ++ { "IEEE_rx_crc", IEEE_R_CRC }, ++ { "IEEE_rx_align", IEEE_R_ALIGN }, ++ { "IEEE_rx_macerr", IEEE_R_MACERR }, ++ { "IEEE_rx_fdxfc", IEEE_R_FDXFC }, ++ { "IEEE_rx_octets_ok", IEEE_R_OCTETS_OK }, ++}; ++ ++static void pfe_eth_fill_stats(struct net_device *dev, struct ethtool_stats *stats, u64 *data) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int i; ++ ++ for (i = 0; i < ARRAY_SIZE(fec_stats); i++) ++ data[i] = readl(priv->EMAC_baseaddr + fec_stats[i].offset); ++} ++ ++static void pfe_eth_gstrings(struct net_device *netdev, ++ u32 stringset, u8 *data) ++{ ++ int i; ++ switch (stringset) { ++ case ETH_SS_STATS: ++ for (i = 0; i < ARRAY_SIZE(fec_stats); i++) ++ memcpy(data + i * ETH_GSTRING_LEN, ++ fec_stats[i].name, ETH_GSTRING_LEN); ++ break; ++ } ++} ++ ++static int pfe_eth_stats_count(struct net_device *dev, int sset) ++{ ++ switch (sset) { ++ case ETH_SS_STATS: ++ return ARRAY_SIZE(fec_stats); ++ default: ++ return -EOPNOTSUPP; ++ } ++} ++ ++/** ++ * pfe_eth_gemac_reglen - Return the length of the register structure. ++ * ++ */ ++static int pfe_eth_gemac_reglen(struct net_device *dev) ++{ ++ printk("%s() \n", __func__); ++ return (sizeof (gemac_regs)/ sizeof(u32)) ; ++} ++ ++/** ++ * pfe_eth_gemac_get_regs - Return the gemac register structure. ++ * ++ */ ++static void pfe_eth_gemac_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *regbuf) ++{ ++ int i; ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ u32 *buf = (u32 *) regbuf; ++ ++ printk("%s() \n", __func__); ++ for (i = 0; i < sizeof (gemac_regs) / sizeof (u32); i++) ++ buf[i] = readl( priv->EMAC_baseaddr + gemac_regs[i] ); ++ ++} ++ ++ ++#endif ++ ++/** ++ * pfe_eth_get_drvinfo - Fills in the drvinfo structure with some basic info ++ * ++ */ ++static void pfe_eth_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *drvinfo) ++{ ++ strncpy(drvinfo->driver, DRV_NAME, COMCERTO_INFOSTR_LEN); ++ strncpy(drvinfo->version, comcerto_eth_driver_version, COMCERTO_INFOSTR_LEN); ++ strncpy(drvinfo->fw_version, "N/A", COMCERTO_INFOSTR_LEN); ++ strncpy(drvinfo->bus_info, "N/A", COMCERTO_INFOSTR_LEN); ++ drvinfo->testinfo_len = 0; ++ drvinfo->regdump_len = 0; ++ drvinfo->eedump_len = 0; ++} ++ ++/** ++ * pfe_eth_set_settings - Used to send commands to PHY. ++ * ++ */ ++ ++static int pfe_eth_set_settings(struct net_device *dev, struct ethtool_cmd *cmd) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ struct phy_device *phydev = priv->phydev; ++ ++ if (NULL == phydev) ++ return -ENODEV; ++ ++ return phy_ethtool_sset(phydev, cmd); ++} ++ ++ ++/** ++ * pfe_eth_getsettings - Return the current settings in the ethtool_cmd structure. ++ * ++ */ ++static int pfe_eth_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ struct phy_device *phydev = priv->phydev; ++ ++ if (NULL == phydev) ++ return -ENODEV; ++ ++ return phy_ethtool_gset(phydev, cmd); ++} ++ ++ ++/** ++ * pfe_eth_get_msglevel - Gets the debug message mask. ++ * ++ */ ++static uint32_t pfe_eth_get_msglevel(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ return priv->msg_enable; ++} ++ ++/** ++ * pfe_eth_set_msglevel - Sets the debug message mask. ++ * ++ */ ++static void pfe_eth_set_msglevel(struct net_device *dev, uint32_t data) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ priv->msg_enable = data; ++} ++ ++#define HIF_RX_COAL_MAX_CLKS (~(1<<31)) ++#define HIF_RX_COAL_CLKS_PER_USEC (pfe->ctrl.sys_clk/1000) ++#define HIF_RX_COAL_MAX_USECS (HIF_RX_COAL_MAX_CLKS/HIF_RX_COAL_CLKS_PER_USEC) ++ ++/** ++ * pfe_eth_set_coalesce - Sets rx interrupt coalescing timer. ++ * ++ */ ++static int pfe_eth_set_coalesce(struct net_device *dev, ++ struct ethtool_coalesce *ec) ++{ ++ if (ec->rx_coalesce_usecs > HIF_RX_COAL_MAX_USECS) ++ return -EINVAL; ++ ++ if (!ec->rx_coalesce_usecs) { ++ writel(0, HIF_INT_COAL); ++ return 0; ++ } ++ ++ writel((ec->rx_coalesce_usecs * HIF_RX_COAL_CLKS_PER_USEC) | HIF_INT_COAL_ENABLE, HIF_INT_COAL); ++ ++ return 0; ++} ++ ++/** ++ * pfe_eth_get_coalesce - Gets rx interrupt coalescing timer value. ++ * ++ */ ++static int pfe_eth_get_coalesce(struct net_device *dev, ++ struct ethtool_coalesce *ec) ++{ ++ int reg_val = readl(HIF_INT_COAL); ++ ++ if (reg_val & HIF_INT_COAL_ENABLE) ++ ec->rx_coalesce_usecs = (reg_val & HIF_RX_COAL_MAX_CLKS) / HIF_RX_COAL_CLKS_PER_USEC; ++ else ++ ec->rx_coalesce_usecs = 0; ++ ++ return 0; ++} ++ ++#if defined(CONFIG_PLATFORM_C2000) ++/** ++ * pfe_eth_pause_rx_enabled - Tests if pause rx is enabled on GEM ++ * ++ */ ++static int pfe_eth_pause_rx_enabled(struct pfe_eth_priv_s *priv) ++{ ++ return (readl(priv->EMAC_baseaddr + EMAC_NETWORK_CONFIG) & EMAC_ENABLE_PAUSE_RX) != 0; ++} ++ ++/** ++ * pfe_eth_set_pauseparam - Sets pause parameters ++ * ++ */ ++static int pfe_eth_set_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ if (epause->rx_pause) ++ { ++ gemac_enable_pause_rx(priv->EMAC_baseaddr); ++ if (priv->phydev) ++ priv->phydev->advertising |= ADVERTISED_Pause | ADVERTISED_Asym_Pause; ++ } ++ else ++ { ++ gemac_disable_pause_rx(priv->EMAC_baseaddr); ++ if (priv->phydev) ++ priv->phydev->advertising &= ~(ADVERTISED_Pause | ADVERTISED_Asym_Pause); ++ } ++ ++ return 0; ++} ++ ++/** ++ * pfe_eth_get_pauseparam - Gets pause parameters ++ * ++ */ ++static void pfe_eth_get_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ epause->autoneg = 0; ++ epause->tx_pause = 0; ++ epause->rx_pause = pfe_eth_pause_rx_enabled(priv); ++} ++ ++/** pfe_eth_get_hash ++ */ ++static int pfe_eth_get_hash(u8 * addr) ++{ ++ u8 temp1,temp2,temp3,temp4,temp5,temp6,temp7,temp8; ++ temp1 = addr[0] & 0x3F ; ++ temp2 = ((addr[0] & 0xC0) >> 6)| ((addr[1] & 0x0F) << 2); ++ temp3 = ((addr[1] & 0xF0) >> 4) | ((addr[2] & 0x03) << 4); ++ temp4 = (addr[2] & 0xFC) >> 2; ++ temp5 = addr[3] & 0x3F; ++ temp6 = ((addr[3] & 0xC0) >> 6) | ((addr[4] & 0x0F) << 2); ++ temp7 = ((addr[4] & 0xF0) >>4 ) | ((addr[5] & 0x03) << 4); ++ temp8 = ((addr[5] &0xFC) >> 2); ++ return (temp1 ^ temp2 ^ temp3 ^ temp4 ^ temp5 ^ temp6 ^ temp7 ^ temp8); ++} ++ ++#else ++ /*TODO Add pause frame support for LS1012A */ ++ ++/** pfe_eth_get_hash ++ */ ++#define HASH_BITS 6 /* #bits in hash */ ++#define CRC32_POLY 0xEDB88320 ++ ++static int pfe_eth_get_hash(u8 * addr) ++{ ++ unsigned int i, bit, data, crc, hash; ++ ++ /* calculate crc32 value of mac address */ ++ crc = 0xffffffff; ++ ++ for (i = 0; i < 6; i++) { ++ data = addr[i]; ++ for (bit = 0; bit < 8; bit++, data >>= 1) { ++ crc = (crc >> 1) ^ ++ (((crc ^ data) & 1) ? CRC32_POLY : 0); ++ } ++ } ++ ++ /* only upper 6 bits (HASH_BITS) are used ++ * which point to specific bit in he hash registers ++ */ ++ hash = (crc >> (32 - HASH_BITS)) & 0x3f; ++ ++ return hash; ++} ++ ++#endif ++ ++struct ethtool_ops pfe_ethtool_ops = { ++ .get_settings = pfe_eth_get_settings, ++ .set_settings = pfe_eth_set_settings, ++ .get_drvinfo = pfe_eth_get_drvinfo, ++ .get_regs_len = pfe_eth_gemac_reglen, ++ .get_regs = pfe_eth_gemac_get_regs, ++ .get_link = ethtool_op_get_link, ++#if defined(CONFIG_PLATFORM_C2000) ++ .get_wol = pfe_eth_get_wol, ++ .set_wol = pfe_eth_set_wol, ++ .set_pauseparam = pfe_eth_set_pauseparam, ++ .get_pauseparam = pfe_eth_get_pauseparam, ++#endif ++ .get_strings = pfe_eth_gstrings, ++ .get_sset_count = pfe_eth_stats_count, ++ .get_ethtool_stats = pfe_eth_fill_stats, ++ .get_msglevel = pfe_eth_get_msglevel, ++ .set_msglevel = pfe_eth_set_msglevel, ++ .set_coalesce = pfe_eth_set_coalesce, ++ .get_coalesce = pfe_eth_get_coalesce, ++}; ++ ++ ++ ++#if defined(CONFIG_PLATFORM_C2000) ++/** pfe_eth_mdio_reset ++ */ ++int pfe_eth_mdio_reset(struct mii_bus *bus) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv; ++ ++ netif_info(priv, hw, priv->dev, "%s\n", __func__); ++ ++#if !defined(CONFIG_PLATFORM_EMULATION) ++ mutex_lock(&bus->mdio_lock); ++ ++ /* Setup the MII Mgmt clock speed */ ++ if (priv->mii_bus) ++ gemac_set_mdc_div(priv->EMAC_baseaddr, priv->mdc_div); ++ ++ /* Reset the management interface */ ++ __raw_writel(__raw_readl(priv->EMAC_baseaddr + EMAC_NETWORK_CONTROL) | EMAC_MDIO_EN, ++ priv->EMAC_baseaddr + EMAC_NETWORK_CONTROL); ++ ++ /* Wait until the bus is free */ ++ while(!(__raw_readl(priv->EMAC_baseaddr + EMAC_NETWORK_STATUS) & EMAC_PHY_IDLE)); ++ ++ mutex_unlock(&bus->mdio_lock); ++#endif ++ ++ return 0; ++} ++ ++ ++/** pfe_eth_gemac_phy_timeout ++ * ++ */ ++static int pfe_eth_gemac_phy_timeout(struct pfe_eth_priv_s *priv, int timeout) ++{ ++ while(!(__raw_readl(priv->EMAC_baseaddr + EMAC_NETWORK_STATUS) & EMAC_PHY_IDLE)) { ++ ++ if (timeout-- <= 0) { ++ return -1; ++ } ++ ++ udelay(10); ++ } ++ ++ return 0; ++} ++ ++ ++/** pfe_eth_mdio_write ++ */ ++static int pfe_eth_mdio_write(struct mii_bus *bus, int mii_id, int regnum, u16 value) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv; ++ u32 write_data; ++ ++#if !defined(CONFIG_PLATFORM_EMULATION) ++ ++ netif_info(priv, hw, priv->dev, "%s: phy %d\n", __func__, mii_id); ++ ++// netif_info(priv, hw, priv->dev, "%s %d %d %x\n", bus->id, mii_id, regnum, value); ++ ++ write_data = 0x50020000; ++ write_data |= ((mii_id << 23) | (regnum << 18) | value); ++ __raw_writel(write_data, priv->EMAC_baseaddr + EMAC_PHY_MANAGEMENT); ++ ++ if (pfe_eth_gemac_phy_timeout(priv, EMAC_MDIO_TIMEOUT)){ ++ netdev_err(priv->dev, "%s: phy MDIO write timeout\n", __func__); ++ return -1; ++ } ++ ++#endif ++ ++ return 0; ++} ++ ++ ++/** pfe_eth_mdio_read ++ */ ++static int pfe_eth_mdio_read(struct mii_bus *bus, int mii_id, int regnum) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv; ++ u16 value = 0; ++ u32 write_data; ++ ++#if !defined(CONFIG_PLATFORM_EMULATION) ++ netif_info(priv, hw, priv->dev, "%s: phy %d\n", __func__, mii_id); ++ ++ write_data = 0x60020000; ++ write_data |= ((mii_id << 23) | (regnum << 18)); ++ ++ __raw_writel(write_data, priv->EMAC_baseaddr + EMAC_PHY_MANAGEMENT); ++ ++ if (pfe_eth_gemac_phy_timeout( priv, EMAC_MDIO_TIMEOUT)) { ++ netdev_err(priv->dev, "%s: phy MDIO read timeout\n", __func__); ++ return -1; ++ } ++ ++ value = __raw_readl(priv->EMAC_baseaddr + EMAC_PHY_MANAGEMENT) & 0xFFFF; ++#endif ++ ++// netif_info(priv, hw, priv->dev, "%s %d %d %x\n", bus->id, mii_id, regnum, value); ++ ++ return value; ++} ++ ++#else ++/** pfe_eth_mdio_reset ++ */ ++int pfe_eth_mdio_reset(struct mii_bus *bus) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv; ++ u32 phy_speed, pclk = 250000000; /*TODO this needs to be checked read from the correct source*/ ++ ++ netif_info(priv, hw, priv->dev, "%s\n", __func__); ++ ++ mutex_lock(&bus->mdio_lock); ++ ++ /* ++ * Set MII speed to 2.5 MHz (= clk_get_rate() / 2 * phy_speed) ++ * ++ * The formula for FEC MDC is 'ref_freq / (MII_SPEED x 2)' while ++ * for ENET-MAC is 'ref_freq / ((MII_SPEED + 1) x 2)'. ++ */ ++ phy_speed = (DIV_ROUND_UP(pclk, 4000000) << EMAC_MII_SPEED_SHIFT); ++ phy_speed |= EMAC_HOLDTIME(0x5); ++ __raw_writel(phy_speed, priv->PHY_baseaddr + EMAC_MII_CTRL_REG); ++ ++ mutex_unlock(&bus->mdio_lock); ++ ++ return 0; ++} ++ ++/** pfe_eth_gemac_phy_timeout ++ * ++ */ ++static int pfe_eth_gemac_phy_timeout(struct pfe_eth_priv_s *priv, int timeout) ++{ ++ while(!(__raw_readl(priv->PHY_baseaddr + EMAC_IEVENT_REG) & EMAC_IEVENT_MII)) { ++ ++ if (timeout-- <= 0) { ++ return -1; ++ } ++ ++ udelay(10); ++ } ++ __raw_writel(EMAC_IEVENT_MII, priv->PHY_baseaddr + EMAC_IEVENT_REG); ++ ++ return 0; ++} ++ ++static int pfe_eth_mdio_mux(u8 muxval) ++{ ++ struct i2c_adapter *a; ++ struct i2c_msg msg; ++ unsigned char buf[2]; ++ int ret; ++ ++ a = i2c_get_adapter(0); ++ if (!a) ++ return -ENODEV; ++ ++ /* set bit 1 (the second bit) of chip at 0x09, register 0x13 */ ++ buf[0] = 0x54; //reg number ++ buf[1] = (muxval << 6)| 0x3; //data ++ msg.addr = 0x66; ++ msg.buf = buf; ++ msg.len = 2; ++ msg.flags = 0; ++ ret = i2c_transfer(a, &msg, 1); ++ i2c_put_adapter(a); ++ if (ret != 1) ++ return -ENODEV; ++ return 0; ++ ++ ++} ++ ++static int pfe_eth_mdio_write(struct mii_bus *bus, int mii_id, int regnum, u16 value) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv; ++ ++ /*FIXME Dirty hack to configure mux */ ++ if(priv->mdio_muxval) { ++ if(mii_id == 0x1) ++ pfe_eth_mdio_mux(0x1); ++ else ++ pfe_eth_mdio_mux(0x2); ++ } ++ ++ /* start a write op */ ++ __raw_writel(EMAC_MII_DATA_ST | EMAC_MII_DATA_OP_WR | ++ EMAC_MII_DATA_PA(mii_id) | EMAC_MII_DATA_RA(regnum) | ++ EMAC_MII_DATA_TA | EMAC_MII_DATA(value), ++ priv->PHY_baseaddr + EMAC_MII_DATA_REG); ++ ++ if (pfe_eth_gemac_phy_timeout(priv, EMAC_MDIO_TIMEOUT)){ ++ netdev_err(priv->dev, "%s: phy MDIO write timeout\n", __func__); ++ return -1; ++ } ++ netif_info(priv, hw, priv->dev, "%s: phy %x reg %x val %x\n", __func__, mii_id, regnum, value); ++ ++ return 0; ++ ++ ++} ++static int pfe_eth_mdio_read(struct mii_bus *bus, int mii_id, int regnum) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)bus->priv; ++ u16 value = 0; ++ ++ /*FIXME Dirty hack to configure mux */ ++ if(priv->mdio_muxval){ ++ if(mii_id == 0x1) ++ pfe_eth_mdio_mux(0x1); ++ else ++ pfe_eth_mdio_mux(0x2); ++ } ++ ++ /* start a read op */ ++ __raw_writel(EMAC_MII_DATA_ST | EMAC_MII_DATA_OP_RD | ++ EMAC_MII_DATA_PA(mii_id) | EMAC_MII_DATA_RA(regnum) | ++ EMAC_MII_DATA_TA, priv->PHY_baseaddr + EMAC_MII_DATA_REG); ++ ++ if (pfe_eth_gemac_phy_timeout( priv, EMAC_MDIO_TIMEOUT)) { ++ netdev_err(priv->dev, "%s: phy MDIO read timeout\n", __func__); ++ return -1; ++ } ++ ++ value = EMAC_MII_DATA(__raw_readl(priv->PHY_baseaddr + EMAC_MII_DATA_REG)); ++ netif_info(priv, hw, priv->dev, "%s: phy %x reg %x val %x\n", __func__, mii_id, regnum, value); ++ return value; ++} ++#endif ++static int pfe_eth_mdio_init(struct pfe_eth_priv_s *priv, struct comcerto_mdio_platform_data *minfo) ++{ ++ struct mii_bus *bus; ++ int rc; ++ ++ netif_info(priv, drv, priv->dev, "%s\n", __func__); ++ printk( "%s\n", __func__); ++ ++#if !defined(CONFIG_PLATFORM_EMULATION) ++ bus = mdiobus_alloc(); ++ if (!bus) { ++ netdev_err(priv->dev, "mdiobus_alloc() failed\n"); ++ rc = -ENOMEM; ++ goto err0; ++ } ++ ++ bus->name = "Comcerto MDIO Bus"; ++ bus->read = &pfe_eth_mdio_read; ++ bus->write = &pfe_eth_mdio_write; ++ bus->reset = &pfe_eth_mdio_reset; ++ snprintf(bus->id, MII_BUS_ID_SIZE, "comcerto-%x", priv->id); ++ bus->priv = priv; ++ ++ bus->phy_mask = minfo->phy_mask; ++ priv->mdc_div = minfo->mdc_div; ++ ++ if (!priv->mdc_div) ++ priv->mdc_div = 64; ++ ++ bus->irq = minfo->irq; ++ ++ bus->parent = priv->pfe->dev; ++ ++ netif_info(priv, drv, priv->dev, "%s: mdc_div: %d, phy_mask: %x \n", __func__, priv->mdc_div, bus->phy_mask); ++ rc = mdiobus_register(bus); ++ if (rc) { ++ netdev_err(priv->dev, "mdiobus_register(%s) failed\n", bus->name); ++ goto err1; ++ } ++ ++ priv->mii_bus = bus; ++ pfe_eth_mdio_reset(bus); ++ ++ return 0; ++ ++err1: ++ mdiobus_free(bus); ++err0: ++ return rc; ++#else ++ return 0; ++#endif ++ ++} ++ ++/** pfe_eth_mdio_exit ++ */ ++static void pfe_eth_mdio_exit(struct mii_bus *bus) ++{ ++ if (!bus) ++ return; ++ ++ netif_info((struct pfe_eth_priv_s *)bus->priv, drv, ((struct pfe_eth_priv_s *)(bus->priv))->dev, "%s\n", __func__); ++ ++ mdiobus_unregister(bus); ++ mdiobus_free(bus); ++} ++ ++#if defined(CONFIG_PLATFORM_C2000) ++/** pfe_get_interface ++ */ ++static phy_interface_t pfe_get_interface(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ u32 mii_mode = priv->einfo->mii_config; ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ if (priv->einfo->gemac_mode & (GEMAC_SW_CONF)) { ++ switch (mii_mode) { ++ case CONFIG_COMCERTO_USE_GMII: ++ return PHY_INTERFACE_MODE_GMII; ++ break; ++ case CONFIG_COMCERTO_USE_RGMII: ++ return PHY_INTERFACE_MODE_RGMII; ++ break; ++ case CONFIG_COMCERTO_USE_RMII: ++ return PHY_INTERFACE_MODE_RMII; ++ break; ++ case CONFIG_COMCERTO_USE_SGMII: ++ return PHY_INTERFACE_MODE_SGMII; ++ break; ++ ++ default : ++ case CONFIG_COMCERTO_USE_MII: ++ return PHY_INTERFACE_MODE_MII; ++ break; ++ ++ } ++ } else { ++ // Bootstrap config read from controller ++ BUG(); ++ return 0; ++ } ++} ++#endif ++ ++/** pfe_get_phydev_speed ++ */ ++static int pfe_get_phydev_speed(struct phy_device *phydev) ++{ ++ switch (phydev->speed) { ++ case 10: ++ return SPEED_10M; ++ case 100: ++ return SPEED_100M; ++ case 1000: ++ default: ++ return SPEED_1000M; ++ } ++ ++} ++ ++/** pfe_set_rgmii_speed ++ */ ++#define RGMIIPCR 0x434 ++/* RGMIIPCR bit definitions*/ ++#define SCFG_RGMIIPCR_EN_AUTO (0x00000008) ++#define SCFG_RGMIIPCR_SETSP_1000M (0x00000004) ++#define SCFG_RGMIIPCR_SETSP_100M (0x00000000) ++#define SCFG_RGMIIPCR_SETSP_10M (0x00000002) ++#define SCFG_RGMIIPCR_SETFD (0x00000001) ++ ++static void pfe_set_rgmii_speed(struct phy_device *phydev) ++{ ++ u32 rgmii_pcr; ++ ++ regmap_read(pfe->scfg, RGMIIPCR, &rgmii_pcr); ++ rgmii_pcr &= ~(SCFG_RGMIIPCR_SETSP_1000M|SCFG_RGMIIPCR_SETSP_10M); ++ ++ switch (phydev->speed) { ++ case 10: ++ rgmii_pcr |= SCFG_RGMIIPCR_SETSP_10M; ++ break; ++ case 1000: ++ rgmii_pcr |= SCFG_RGMIIPCR_SETSP_1000M; ++ break; ++ case 100: ++ default: ++ /* Default is 100M */ ++ break; ++ } ++ regmap_write(pfe->scfg, RGMIIPCR, rgmii_pcr); ++ ++ ++} ++/** pfe_get_phydev_duplex ++ */ ++static int pfe_get_phydev_duplex(struct phy_device *phydev) ++{ ++ //return ( phydev->duplex == DUPLEX_HALF ) ? DUP_HALF:DUP_FULL ; ++ return DUPLEX_FULL; ++} ++ ++/** pfe_eth_adjust_link ++ */ ++static void pfe_eth_adjust_link(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ unsigned long flags; ++ struct phy_device *phydev = priv->phydev; ++ int new_state = 0; ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ spin_lock_irqsave(&priv->lock, flags); ++ if (phydev->link) { ++ /* Now we make sure that we can be in full duplex mode. ++ * If not, we operate in half-duplex mode. */ ++ if (phydev->duplex != priv->oldduplex) { ++ new_state = 1; ++ gemac_set_duplex(priv->EMAC_baseaddr, pfe_get_phydev_duplex(phydev)); ++ priv->oldduplex = phydev->duplex; ++ } ++ ++ if (phydev->speed != priv->oldspeed) { ++ new_state = 1; ++ gemac_set_speed(priv->EMAC_baseaddr, pfe_get_phydev_speed(phydev)); ++ if(priv->einfo->mii_config == PHY_INTERFACE_MODE_RGMII) ++ pfe_set_rgmii_speed(phydev); ++ priv->oldspeed = phydev->speed; ++ } ++ ++ if (!priv->oldlink) { ++ new_state = 1; ++ priv->oldlink = 1; ++ } ++ ++ } else if (priv->oldlink) { ++ new_state = 1; ++ priv->oldlink = 0; ++ priv->oldspeed = 0; ++ priv->oldduplex = -1; ++ } ++ ++ if (new_state && netif_msg_link(priv)) ++ phy_print_status(phydev); ++ ++ spin_unlock_irqrestore(&priv->lock, flags); ++} ++ ++ ++/** pfe_phy_exit ++ */ ++static void pfe_phy_exit(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ phy_disconnect(priv->phydev); ++ priv->phydev = NULL; ++} ++ ++/** pfe_eth_stop ++ */ ++static void pfe_eth_stop( struct net_device *dev , int wake) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ if (wake) ++ gemac_tx_disable(priv->EMAC_baseaddr); ++ else { ++ gemac_disable(priv->EMAC_baseaddr); ++ gpi_disable(priv->GPI_baseaddr); ++ ++ if (priv->phydev) ++ phy_stop(priv->phydev); ++ } ++} ++ ++/** pfe_eth_start ++ */ ++static int pfe_eth_start( struct pfe_eth_priv_s *priv ) ++{ ++ netif_info(priv, drv, priv->dev, "%s\n", __func__); ++ ++ if (priv->phydev) ++ phy_start(priv->phydev); ++ ++ gpi_enable(priv->GPI_baseaddr); ++ gemac_enable(priv->EMAC_baseaddr); ++ ++ return 0; ++} ++ ++/*Configure on chip serdes through mdio ++ * Is there any better way to do this? */ ++static void ls1012a_configure_serdes(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = pfe->eth.eth_priv[0]; // FIXME This will not work for EMAC2 as SGMII ++ /*int value,sgmii_2500=0; */ ++ struct mii_bus *bus = priv->mii_bus; ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ /* PCS configuration done with corresponding GEMAC */ ++ ++ pfe_eth_mdio_read(bus, 0, 0); ++ pfe_eth_mdio_read(bus, 0, 1); ++#if 1 ++ /*These settings taken from validtion team */ ++ pfe_eth_mdio_write(bus, 0, 0x0, 0x8000); ++ pfe_eth_mdio_write(bus, 0, 0x14, 0xb); ++ pfe_eth_mdio_write(bus, 0, 0x4, 0x1a1); ++ pfe_eth_mdio_write(bus, 0, 0x12, 0x400); ++ pfe_eth_mdio_write(bus, 0, 0x13, 0x0); ++ pfe_eth_mdio_write(bus, 0, 0x0, 0x1140); ++ return; ++#else ++ /*Reset serdes */ ++ pfe_eth_mdio_write(bus, 0, 0x0, 0x8000); ++ ++ /* SGMII IF mode + AN enable only for 1G SGMII, not for 2.5G */ ++ value = PHY_SGMII_IF_MODE_SGMII; ++ if (!sgmii_2500) ++ value |= PHY_SGMII_IF_MODE_AN; ++ ++ pfe_eth_mdio_write(bus, 0, 0x14, value); ++ ++ /* Dev ability according to SGMII specification */ ++ value = PHY_SGMII_DEV_ABILITY_SGMII; ++ pfe_eth_mdio_write(bus, 0, 0x4, value); ++ ++ //These values taken from validation team ++ pfe_eth_mdio_write(bus, 0, 0x13, 0x0); ++ pfe_eth_mdio_write(bus, 0, 0x12, 0x400); ++ ++ /* Restart AN */ ++ value = PHY_SGMII_CR_DEF_VAL; ++ if (!sgmii_2500) ++ value |= PHY_SGMII_CR_RESET_AN; ++ pfe_eth_mdio_write(bus, 0, 0, value); ++ ++#endif ++} ++ ++/** pfe_phy_init ++ * ++ */ ++static int pfe_phy_init(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ struct phy_device *phydev; ++ char phy_id[MII_BUS_ID_SIZE + 3]; ++ char bus_id[MII_BUS_ID_SIZE]; ++ phy_interface_t interface; ++ ++ priv->oldlink = 0; ++ priv->oldspeed = 0; ++ priv->oldduplex = -1; ++ ++ snprintf(bus_id, MII_BUS_ID_SIZE, "comcerto-%d", 0); ++ snprintf(phy_id, MII_BUS_ID_SIZE + 3, PHY_ID_FMT, bus_id, priv->einfo->phy_id); ++ ++ netif_info(priv, drv, dev, "%s: %s\n", __func__, phy_id); ++#if defined(CONFIG_PLATFORM_C2000) ++ interface = pfe_get_interface(dev); ++#else ++ interface = priv->einfo->mii_config; ++ if(interface == PHY_INTERFACE_MODE_SGMII) { ++ /*Configure SGMII PCS */ ++ if(pfe->scfg) { ++ /*Config MDIO from serdes */ ++ regmap_write(pfe->scfg, 0x484, 0x00000000); ++ } ++ ls1012a_configure_serdes(dev); ++ } ++ ++ if(pfe->scfg) { ++ /*Config MDIO from PAD */ ++ regmap_write(pfe->scfg, 0x484, 0x80000000); ++ } ++#endif ++ ++ ++ priv->oldlink = 0; ++ priv->oldspeed = 0; ++ priv->oldduplex = -1; ++ ++ printk("%s interface %x \n", __func__, interface); ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,9,0) ++ phydev = phy_connect(dev, phy_id, &pfe_eth_adjust_link, interface); ++#else ++ phydev = phy_connect(dev, phy_id, &pfe_eth_adjust_link, 0, interface); ++#endif ++ ++ if (IS_ERR(phydev)) { ++ netdev_err(dev, "phy_connect() failed\n"); ++ return PTR_ERR(phydev); ++ } ++ ++ priv->phydev = phydev; ++ phydev->irq = PHY_POLL; ++ ++#if defined(CONFIG_PLATFORM_C2000) ++ /* Pause frame support */ ++ phydev->supported |= SUPPORTED_Pause | SUPPORTED_Asym_Pause; ++ if (pfe_eth_pause_rx_enabled(priv)) ++ phydev->advertising |= ADVERTISED_Pause | ADVERTISED_Asym_Pause; ++ else ++ phydev->advertising &= ~(ADVERTISED_Pause | ADVERTISED_Asym_Pause); ++#else ++ /*TODO Add pause frame support for LS1012A */ ++#endif ++ ++ return 0; ++} ++ ++/** pfe_gemac_init ++ */ ++static int pfe_gemac_init(struct pfe_eth_priv_s *priv) ++{ ++ GEMAC_CFG cfg; ++ ++ netif_info(priv, ifup, priv->dev, "%s\n", __func__); ++ ++ /* software config */ ++ /* MII interface mode selection */ ++ switch (priv->einfo->mii_config) { ++ case CONFIG_COMCERTO_USE_GMII: ++ cfg.mode = GMII; ++ break; ++ ++ case CONFIG_COMCERTO_USE_MII: ++ cfg.mode = MII; ++ break; ++ ++ case CONFIG_COMCERTO_USE_RGMII: ++ cfg.mode = RGMII; ++ break; ++ ++ case CONFIG_COMCERTO_USE_RMII: ++ cfg.mode = RMII; ++ break; ++ ++ case CONFIG_COMCERTO_USE_SGMII: ++ cfg.mode = SGMII; ++ break; ++ ++ default: ++ cfg.mode = RGMII; ++ } ++ ++ /* Speed selection */ ++ switch (priv->einfo->gemac_mode & GEMAC_SW_SPEED_1G ) { ++ case GEMAC_SW_SPEED_1G: ++ cfg.speed = SPEED_1000M; ++ break; ++ ++ case GEMAC_SW_SPEED_100M: ++ cfg.speed = SPEED_100M; ++ break; ++ ++ case GEMAC_SW_SPEED_10M: ++ cfg.speed = SPEED_10M; ++ break; ++ ++ default: ++ cfg.speed = SPEED_1000M; ++ } ++ ++ /* Duplex selection */ ++ cfg.duplex = ( priv->einfo->gemac_mode & GEMAC_SW_FULL_DUPLEX ) ? DUPLEX_FULL : DUPLEX_HALF; ++ ++ gemac_set_config( priv->EMAC_baseaddr, &cfg); ++ gemac_allow_broadcast( priv->EMAC_baseaddr ); ++ gemac_disable_unicast( priv->EMAC_baseaddr ); ++ gemac_disable_multicast( priv->EMAC_baseaddr ); ++ gemac_disable_fcs_rx( priv->EMAC_baseaddr ); ++ gemac_enable_1536_rx( priv->EMAC_baseaddr ); ++ gemac_enable_rx_jmb( priv->EMAC_baseaddr ); ++ gemac_enable_stacked_vlan( priv->EMAC_baseaddr ); ++ gemac_enable_pause_rx( priv->EMAC_baseaddr ); ++ gemac_set_bus_width(priv->EMAC_baseaddr, 64); ++ /*TODO just for testing remove it later */ ++ gemac_enable_copy_all(priv->EMAC_baseaddr); ++ ++ /*GEM will perform checksum verifications*/ ++ if (priv->dev->features & NETIF_F_RXCSUM) ++ gemac_enable_rx_checksum_offload(priv->EMAC_baseaddr); ++ else ++ gemac_disable_rx_checksum_offload(priv->EMAC_baseaddr); ++ ++ return 0; ++} ++ ++/** pfe_eth_event_handler ++ */ ++static int pfe_eth_event_handler(void *data, int event, int qno) ++{ ++ struct pfe_eth_priv_s *priv = data; ++ ++ switch (event) { ++ case EVENT_RX_PKT_IND: ++ ++ if (qno == 0) { ++ if (napi_schedule_prep(&priv->high_napi)) { ++ netif_info(priv, intr, priv->dev, "%s: schedule high prio poll\n", __func__); ++ ++#ifdef PFE_ETH_NAPI_STATS ++ priv->napi_counters[NAPI_SCHED_COUNT]++; ++#endif ++ ++ __napi_schedule(&priv->high_napi); ++ } ++ } ++ else if (qno == 1) { ++ if (napi_schedule_prep(&priv->low_napi)) { ++ netif_info(priv, intr, priv->dev, "%s: schedule low prio poll\n", __func__); ++ ++#ifdef PFE_ETH_NAPI_STATS ++ priv->napi_counters[NAPI_SCHED_COUNT]++; ++#endif ++ __napi_schedule(&priv->low_napi); ++ } ++ } ++ else if (qno == 2) { ++ if (napi_schedule_prep(&priv->lro_napi)) { ++ netif_info(priv, intr, priv->dev, "%s: schedule lro prio poll\n", __func__); ++ ++#ifdef PFE_ETH_NAPI_STATS ++ priv->napi_counters[NAPI_SCHED_COUNT]++; ++#endif ++ __napi_schedule(&priv->lro_napi); ++ } ++ } ++ ++ break; ++ ++ case EVENT_TXDONE_IND: ++ case EVENT_HIGH_RX_WM: ++ default: ++ break; ++ } ++ ++ return 0; ++} ++ ++/** pfe_eth_open ++ */ ++static int pfe_eth_open(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ struct hif_client_s *client; ++ int rc; ++ ++ netif_info(priv, ifup, dev, "%s\n", __func__); ++ ++ /* Register client driver with HIF */ ++ client = &priv->client; ++ memset(client, 0, sizeof(*client)); ++ client->id = PFE_CL_GEM0 + priv->id; ++ client->tx_qn = emac_txq_cnt; ++ client->rx_qn = EMAC_RXQ_CNT; ++ client->priv = priv; ++ client->pfe = priv->pfe; ++ client->event_handler = pfe_eth_event_handler; ++ ++ /* FIXME : For now hif lib sets all tx and rx queues to same size */ ++ client->tx_qsize = EMAC_TXQ_DEPTH; ++ client->rx_qsize = EMAC_RXQ_DEPTH; ++ ++ if ((rc = hif_lib_client_register(client))) { ++ netdev_err(dev, "%s: hif_lib_client_register(%d) failed\n", __func__, client->id); ++ goto err0; ++ } ++ ++ netif_info(priv, drv, dev, "%s: registered client: %p\n", __func__, client); ++ ++#if defined(CONFIG_PLATFORM_C2000) ++ /* Enable gemac tx clock */ ++ clk_enable(priv->gemtx_clk); ++#endif ++ ++ pfe_gemac_init(priv); ++ ++ if (!is_valid_ether_addr(dev->dev_addr)) { ++ netdev_err(dev, "%s: invalid MAC address\n", __func__); ++ rc = -EADDRNOTAVAIL; ++ goto err1; ++ } ++ ++ gemac_set_laddrN( priv->EMAC_baseaddr, ( MAC_ADDR *)dev->dev_addr, 1 ); ++ ++ napi_enable(&priv->high_napi); ++ napi_enable(&priv->low_napi); ++ napi_enable(&priv->lro_napi); ++ ++ rc = pfe_eth_start(priv); ++ ++ netif_tx_wake_all_queues(dev); ++ ++ //pfe_ctrl_set_eth_state(priv->id, 1, dev->dev_addr); ++ ++ priv->tx_timer.expires = jiffies + ( COMCERTO_TX_RECOVERY_TIMEOUT_MS * HZ )/1000; ++ add_timer(&priv->tx_timer); ++ ++ return rc; ++ ++err1: ++ hif_lib_client_unregister(&priv->client); ++#if defined(CONFIG_PLATFORM_C2000) ++ clk_disable(priv->gemtx_clk); ++#endif ++ ++err0: ++ return rc; ++} ++/* ++ * pfe_eth_shutdown ++ */ ++int pfe_eth_shutdown( struct net_device *dev, int wake) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int i, qstatus; ++ unsigned long next_poll = jiffies + 1, end = jiffies + (TX_POLL_TIMEOUT_MS * HZ) / 1000; ++ int tx_pkts, prv_tx_pkts; ++ ++ netif_info(priv, ifdown, dev, "%s\n", __func__); ++ ++ del_timer_sync(&priv->tx_timer); ++ ++ for(i = 0; i < emac_txq_cnt; i++) ++ hrtimer_cancel(&priv->fast_tx_timeout[i].timer); ++ ++ netif_tx_stop_all_queues(dev); ++ ++ do { ++ tx_pkts = 0; ++ pfe_eth_flush_tx(priv, 1); ++ ++ for (i = 0; i < emac_txq_cnt; i++) ++ tx_pkts += hif_lib_tx_pending(&priv->client, i); ++ ++ if (tx_pkts) { ++ /*Don't wait forever, break if we cross max timeout */ ++ if (time_after(jiffies, end)) { ++ printk(KERN_ERR "(%s)Tx is not complete after %dmsec\n", dev->name, TX_POLL_TIMEOUT_MS); ++ break; ++ } ++ ++ printk("%s : (%s) Waiting for tx packets to free. Pending tx pkts = %d.\n", __func__, dev->name, tx_pkts); ++ if (need_resched()) ++ schedule(); ++ } ++ ++ } while(tx_pkts); ++ ++ end = jiffies + (TX_POLL_TIMEOUT_MS * HZ) / 1000; ++ /*Disable transmit in PFE before disabling GEMAC */ ++ //pfe_ctrl_set_eth_state(priv->id, 0, NULL); ++ ++ prv_tx_pkts = tmu_pkts_processed(priv->id); ++ /*Wait till TMU transmits all pending packets ++ * poll tmu_qstatus and pkts processed by TMU for every 10ms ++ * Consider TMU is busy, If we see TMU qeueu pending or any packets processed by TMU ++ */ ++ while(1) { ++ ++ if (time_after(jiffies, next_poll)) { ++ ++ tx_pkts = tmu_pkts_processed(priv->id); ++ qstatus = tmu_qstatus(priv->id) & 0x7ffff; ++ ++ if(!qstatus && (tx_pkts == prv_tx_pkts)) { ++ break; ++ } ++ /*Don't wait forever, break if we cross max timeout(TX_POLL_TIMEOUT_MS) */ ++ if (time_after(jiffies, end)) { ++ printk(KERN_ERR "TMU%d is busy after %dmsec\n", priv->id, TX_POLL_TIMEOUT_MS); ++ break; ++ } ++ prv_tx_pkts = tx_pkts; ++ next_poll++; ++ } ++ if (need_resched()) ++ schedule(); ++ ++ ++ } ++ /* Wait for some more time to complete transmitting packet if any */ ++ next_poll = jiffies + 1; ++ while(1) { ++ if (time_after(jiffies, next_poll)) ++ break; ++ if (need_resched()) ++ schedule(); ++ } ++ ++ pfe_eth_stop(dev, wake); ++ ++ napi_disable(&priv->lro_napi); ++ napi_disable(&priv->low_napi); ++ napi_disable(&priv->high_napi); ++ ++#if defined(CONFIG_PLATFORM_C2000) ++ /* Disable gemac tx clock */ ++ clk_disable(priv->gemtx_clk); ++#endif ++ ++ hif_lib_client_unregister(&priv->client); ++ ++ return 0; ++} ++ ++/* pfe_eth_close ++ * ++ */ ++static int pfe_eth_close( struct net_device *dev ) ++{ ++ pfe_eth_shutdown(dev, 0); ++ ++ return 0; ++} ++ ++/* pfe_eth_suspend ++ * ++ * return value : 1 if netdevice is configured to wakeup system ++ * 0 otherwise ++ */ ++int pfe_eth_suspend(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int retval = 0; ++ ++ if (priv->wol) { ++ gemac_set_wol(priv->EMAC_baseaddr, priv->wol); ++ retval = 1; ++ } ++ pfe_eth_shutdown(dev, priv->wol); ++ ++ return retval; ++} ++ ++/** pfe_eth_resume ++ * ++ */ ++int pfe_eth_resume(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ if (priv->wol) ++ gemac_set_wol(priv->EMAC_baseaddr, 0); ++ ++ return pfe_eth_open(dev); ++} ++ ++#if defined(CONFIG_PLATFORM_C2000) ++/** pfe_eth_set_device_wakeup ++ * ++ * Called when a netdevice changes its wol status. ++ * Scans state of all interfaces and updae PFE device ++ * wakeable state ++ */ ++static void pfe_eth_set_device_wakeup(struct pfe *pfe) ++{ ++ int i; ++ int wake = 0; ++ ++ for(i = 0; i < NUM_GEMAC_SUPPORT; i++) ++ wake |= pfe->eth.eth_priv[i]->wol; ++ ++ device_set_wakeup_enable(pfe->dev, wake); ++ //TODO Find correct IRQ mapping. ++ //TODO interface with PMU ++ //int irq_set_irq_wake(unsigned int irq, unsigned int on) ++} ++#endif ++/** pfe_eth_get_queuenum ++ * ++ */ ++static int pfe_eth_get_queuenum( struct pfe_eth_priv_s *priv, struct sk_buff *skb ) ++{ ++ int queuenum = 0; ++ unsigned long flags; ++ ++ /* Get the Fast Path queue number */ ++ /* Use conntrack mark (if conntrack exists), then packet mark (if any), then fallback to default */ ++#if defined(CONFIG_IP_NF_CONNTRACK_MARK) || defined(CONFIG_NF_CONNTRACK_MARK) ++ if (skb->nfct) { ++ enum ip_conntrack_info cinfo; ++ struct nf_conn *ct; ++ ct = nf_ct_get(skb, &cinfo); ++ ++ if (ct) { ++ u_int32_t connmark; ++ connmark = ct->mark; ++ ++ if ((connmark & 0x80000000) && priv->id != 0) ++ connmark >>= 16; ++ ++ queuenum = connmark & EMAC_QUEUENUM_MASK; ++ } ++ } ++ else /* continued after #endif ... */ ++#endif ++ if (skb->mark) ++ queuenum = skb->mark & EMAC_QUEUENUM_MASK; ++ else { ++ spin_lock_irqsave(&priv->lock, flags); ++ queuenum = priv->default_priority & EMAC_QUEUENUM_MASK; ++ spin_unlock_irqrestore(&priv->lock, flags); ++ } ++ ++ return queuenum; ++} ++ ++ ++ ++/** pfe_eth_might_stop_tx ++ * ++ */ ++static int pfe_eth_might_stop_tx(struct pfe_eth_priv_s *priv, int queuenum, struct netdev_queue *tx_queue, unsigned int n_desc, unsigned int n_segs) ++{ ++ int tried = 0; ++ ktime_t kt; ++ ++try_again: ++ if (unlikely((__hif_tx_avail(&pfe->hif) < n_desc) ++ || (hif_lib_tx_avail(&priv->client, queuenum) < n_desc) ++ || (hif_lib_tx_credit_avail(pfe, priv->id, queuenum) < n_segs))) { ++ ++ if (!tried) { ++ hif_tx_unlock(&pfe->hif); ++ pfe_eth_flush_txQ(priv, queuenum, 1, n_desc); ++ hif_lib_update_credit(&priv->client, queuenum); ++ tried = 1; ++ hif_tx_lock(&pfe->hif); ++ goto try_again; ++ } ++#ifdef PFE_ETH_TX_STATS ++ if (__hif_tx_avail(&pfe->hif) < n_desc) ++ priv->stop_queue_hif[queuenum]++; ++ else if (hif_lib_tx_avail(&priv->client, queuenum) < n_desc) { ++ priv->stop_queue_hif_client[queuenum]++; ++ } ++ else if (hif_lib_tx_credit_avail(pfe, priv->id, queuenum) < n_segs) { ++ priv->stop_queue_credit[queuenum]++; ++ } ++ priv->stop_queue_total[queuenum]++; ++#endif ++ netif_tx_stop_queue(tx_queue); ++ ++ kt = ktime_set(0, COMCERTO_TX_FAST_RECOVERY_TIMEOUT_MS * NSEC_PER_MSEC); ++ hrtimer_start(&priv->fast_tx_timeout[queuenum].timer, kt, HRTIMER_MODE_REL); ++ return -1; ++ } ++ else { ++ return 0; ++ } ++} ++ ++#define SA_MAX_OP 2 ++/** pfe_hif_send_packet ++ * ++ * At this level if TX fails we drop the packet ++ */ ++static void pfe_hif_send_packet( struct sk_buff *skb, struct pfe_eth_priv_s *priv, int queuenum) ++{ ++ struct skb_shared_info *sh = skb_shinfo(skb); ++ unsigned int nr_frags; ++ u32 ctrl = 0; ++ ++ netif_info(priv, tx_queued, priv->dev, "%s\n", __func__); ++ ++ if (skb_is_gso(skb)) { ++ priv->stats.tx_dropped++; ++ return; ++ } ++ ++ if (skb->ip_summed == CHECKSUM_PARTIAL) { ++ if (skb->len > 1522) { ++ skb->ip_summed = 0; ++ ctrl = 0; ++ ++ if (pfe_compute_csum(skb)){ ++ kfree_skb(skb); ++ return; ++ } ++ } ++ else ++ ctrl = HIF_CTRL_TX_CHECKSUM; ++ } ++ ++ nr_frags = sh->nr_frags; ++ ++ if (nr_frags) { ++ skb_frag_t *f; ++ int i; ++ ++ __hif_lib_xmit_pkt(&priv->client, queuenum, skb->data, skb_headlen(skb), ctrl, HIF_FIRST_BUFFER, skb); ++ ++ for (i = 0; i < nr_frags - 1; i++) { ++ f = &sh->frags[i]; ++ __hif_lib_xmit_pkt(&priv->client, queuenum, skb_frag_address(f), skb_frag_size(f), 0x0, 0x0, skb); ++ } ++ ++ f = &sh->frags[i]; ++ ++ __hif_lib_xmit_pkt(&priv->client, queuenum, skb_frag_address(f), skb_frag_size(f), 0x0, HIF_LAST_BUFFER|HIF_DATA_VALID, skb); ++ ++ netif_info(priv, tx_queued, priv->dev, "%s: pkt sent successfully skb:%p nr_frags:%d len:%d\n", __func__, skb, nr_frags, skb->len); ++ } ++ else { ++ __hif_lib_xmit_pkt(&priv->client, queuenum, skb->data, skb->len, ctrl, HIF_FIRST_BUFFER | HIF_LAST_BUFFER | HIF_DATA_VALID, skb); ++ netif_info(priv, tx_queued, priv->dev, "%s: pkt sent successfully skb:%p len:%d\n", __func__, skb, skb->len); ++ } ++ hif_tx_dma_start(); ++ priv->stats.tx_packets++; ++ priv->stats.tx_bytes += skb->len; ++ hif_lib_tx_credit_use(pfe, priv->id, queuenum, 1); ++} ++ ++/** pfe_eth_flush_txQ ++ */ ++static void pfe_eth_flush_txQ(struct pfe_eth_priv_s *priv, int txQ_num, int from_tx, int n_desc) ++{ ++ struct sk_buff *skb; ++ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->dev, txQ_num); ++ int count = max(TX_FREE_MAX_COUNT, n_desc); ++ unsigned int flags; ++ ++ netif_info(priv, tx_done, priv->dev, "%s\n", __func__); ++ ++ if (!from_tx) ++ __netif_tx_lock_bh(tx_queue); ++ ++ /* Clean HIF and client queue */ ++ while (count && (skb = hif_lib_tx_get_next_complete(&priv->client, txQ_num, &flags, count))) { ++ ++ /* FIXME : Invalid data can be skipped in hif_lib itself */ ++ if (flags & HIF_DATA_VALID) { ++ dev_kfree_skb_any(skb); ++ ++ } ++ // When called from the timer, flush all descriptors ++ if (from_tx) ++ count--; ++ } ++ ++ if (!from_tx) ++ __netif_tx_unlock_bh(tx_queue); ++} ++ ++/** pfe_eth_flush_tx ++ */ ++static void pfe_eth_flush_tx(struct pfe_eth_priv_s *priv, int force) ++{ ++ int ii; ++ ++ netif_info(priv, tx_done, priv->dev, "%s\n", __func__); ++ ++ for (ii = 0; ii < emac_txq_cnt; ii++) { ++ if (force || (time_after(jiffies, priv->client.tx_q[ii].jiffies_last_packet + (COMCERTO_TX_RECOVERY_TIMEOUT_MS * HZ)/1000))) { ++ pfe_eth_flush_txQ(priv, ii, 0, 0); //We will release everything we can based on from_tx param, so the count param can be set to any value ++ hif_lib_update_credit(&priv->client, ii); ++ } ++ } ++} ++ ++void pfe_tx_get_req_desc(struct sk_buff *skb, unsigned int *n_desc, unsigned int *n_segs) ++{ ++ struct skb_shared_info *sh = skb_shinfo(skb); ++ ++ // Scattered data ++ if (sh->nr_frags) { ++ *n_desc = sh->nr_frags + 1; ++ *n_segs = 1; ++ } ++ // Regular case ++ else { ++ *n_desc = 1; ++ *n_segs = 1; ++ } ++ return; ++} ++ ++/** pfe_eth_send_packet ++ */ ++static int pfe_eth_send_packet(struct sk_buff *skb, struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int txQ_num = skb_get_queue_mapping(skb); ++ int n_desc, n_segs, count; ++ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->dev, txQ_num); ++ ++ netif_info(priv, tx_queued, dev, "%s\n", __func__); ++ ++ if ((!skb_is_gso(skb)) && (skb_headroom(skb) < (PFE_PKT_HEADER_SZ + sizeof(unsigned long)))) { ++ ++ netif_warn(priv, tx_err, priv->dev, "%s: copying skb\n", __func__); ++ ++ if (pskb_expand_head(skb, (PFE_PKT_HEADER_SZ + sizeof(unsigned long)), 0, GFP_ATOMIC)) { ++ /* No need to re-transmit, no way to recover*/ ++ kfree_skb(skb); ++ priv->stats.tx_dropped++; ++ return NETDEV_TX_OK; ++ } ++ } ++ ++ pfe_tx_get_req_desc(skb, &n_desc, &n_segs); ++ ++ hif_tx_lock(&pfe->hif); ++ if(unlikely(pfe_eth_might_stop_tx(priv, txQ_num, tx_queue, n_desc, n_segs))) { ++#ifdef PFE_ETH_TX_STATS ++ if(priv->was_stopped[txQ_num]) { ++ priv->clean_fail[txQ_num]++; ++ priv->was_stopped[txQ_num] = 0; ++ } ++#endif ++ hif_tx_unlock(&pfe->hif); ++ return NETDEV_TX_BUSY; ++ } ++ ++ pfe_hif_send_packet(skb, priv, txQ_num); ++ ++ hif_tx_unlock(&pfe->hif); ++ ++ dev->trans_start = jiffies; ++ ++ // Recycle buffers if a socket's send buffer becomes half full or if the HIF client queue starts filling up ++ if (((count = (hif_lib_tx_pending(&priv->client, txQ_num) - HIF_CL_TX_FLUSH_MARK)) > 0) ++ || (skb->sk && ((sk_wmem_alloc_get(skb->sk) << 1) > skb->sk->sk_sndbuf))) ++ pfe_eth_flush_txQ(priv, txQ_num, 1, count); ++ ++#ifdef PFE_ETH_TX_STATS ++ priv->was_stopped[txQ_num] = 0; ++#endif ++ ++ return NETDEV_TX_OK; ++} ++ ++/** pfe_eth_select_queue ++ * ++ */ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,14,0) ++static u16 pfe_eth_select_queue( struct net_device *dev, struct sk_buff *skb, ++ void *accel_priv, select_queue_fallback_t fallback) ++#else ++static u16 pfe_eth_select_queue( struct net_device *dev, struct sk_buff *skb ) ++#endif ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ return pfe_eth_get_queuenum(priv, skb); ++} ++ ++ ++/** pfe_eth_get_stats ++ */ ++static struct net_device_stats *pfe_eth_get_stats(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ return &priv->stats; ++} ++ ++ ++/** pfe_eth_change_mtu ++ */ ++static int pfe_eth_change_mtu(struct net_device *dev, int new_mtu) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int oldsize = dev->mtu ; ++ int frame_size = new_mtu + ETH_HLEN +4; ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ if ((frame_size < 64) || (frame_size > JUMBO_FRAME_SIZE)) { ++ netif_err(priv, drv, dev, "Invalid MTU setting\n"); ++ return -EINVAL; ++ } ++ ++ if ((new_mtu > 1500) && (dev->features & NETIF_F_TSO)) ++ { ++ priv->usr_features = dev->features; ++ if (dev->features & NETIF_F_TSO) ++ { ++ netdev_err(dev, "MTU cannot be set to more than 1500 while TSO is enabled. disabling TSO.\n"); ++ dev->features &= ~(NETIF_F_TSO); ++ } ++ } ++ else if ((dev->mtu > 1500) && (new_mtu <= 1500)) ++ { ++ if (priv->usr_features & NETIF_F_TSO) ++ { ++ priv->usr_features &= ~(NETIF_F_TSO); ++ dev->features |= NETIF_F_TSO; ++ netdev_err(dev, "MTU is <= 1500, Enabling TSO feature.\n"); ++ } ++ } ++ ++ /* Only stop and start the controller if it isn't already ++ * stopped, and we changed something */ ++ if ((oldsize != new_mtu) && (dev->flags & IFF_UP)){ ++ netdev_err(dev, "Can not change MTU - fast_path must be disabled and ifconfig down must be issued first\n"); ++ ++ return -EINVAL; ++ } ++ ++ dev->mtu = new_mtu; ++ ++ return 0; ++} ++ ++/** pfe_eth_set_mac_address ++ */ ++static int pfe_eth_set_mac_address(struct net_device *dev, void *addr) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ struct sockaddr *sa = addr; ++ ++ netif_info(priv, drv, dev, "%s\n", __func__); ++ ++ if (!is_valid_ether_addr(sa->sa_data)) ++ return -EADDRNOTAVAIL; ++ ++ memcpy(dev->dev_addr, sa->sa_data, ETH_ALEN); ++ ++ gemac_set_laddrN(priv->EMAC_baseaddr, (MAC_ADDR *)dev->dev_addr, 1); ++ ++ return 0; ++ ++} ++ ++/** pfe_eth_enet_addr_byte_mac ++ */ ++int pfe_eth_enet_addr_byte_mac(u8 * enet_byte_addr, MAC_ADDR *enet_addr) ++{ ++ if ((enet_byte_addr == NULL) || (enet_addr == NULL)) ++ { ++ return -1; ++ } ++ else ++ { ++ enet_addr->bottom = enet_byte_addr[0] | ++ (enet_byte_addr[1] << 8) | ++ (enet_byte_addr[2] << 16) | ++ (enet_byte_addr[3] << 24); ++ enet_addr->top = enet_byte_addr[4] | ++ (enet_byte_addr[5] << 8); ++ return 0; ++ } ++} ++ ++ ++ ++/** pfe_eth_set_multi ++ */ ++static void pfe_eth_set_multi(struct net_device *dev) ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ MAC_ADDR hash_addr; /* hash register structure */ ++ MAC_ADDR spec_addr; /* specific mac address register structure */ ++ int result; /* index into hash register to set.. */ ++ int uc_count = 0; ++ struct netdev_hw_addr *ha; ++ ++ if (dev->flags & IFF_PROMISC) { ++ netif_info(priv, drv, dev, "entering promiscuous mode\n"); ++ ++ priv->promisc = 1; ++ gemac_enable_copy_all(priv->EMAC_baseaddr); ++ } else { ++ priv->promisc = 0; ++ gemac_disable_copy_all(priv->EMAC_baseaddr); ++ } ++ ++ /* Enable broadcast frame reception if required. */ ++ if (dev->flags & IFF_BROADCAST) { ++ gemac_allow_broadcast(priv->EMAC_baseaddr); ++ } else { ++ netif_info(priv, drv, dev, "disabling broadcast frame reception\n"); ++ ++ gemac_no_broadcast(priv->EMAC_baseaddr); ++ } ++ ++ if (dev->flags & IFF_ALLMULTI) { ++ /* Set the hash to rx all multicast frames */ ++ hash_addr.bottom = 0xFFFFFFFF; ++ hash_addr.top = 0xFFFFFFFF; ++ gemac_set_hash(priv->EMAC_baseaddr, &hash_addr); ++ gemac_enable_multicast(priv->EMAC_baseaddr); ++ netdev_for_each_uc_addr(ha, dev) { ++ if(uc_count >= MAX_UC_SPEC_ADDR_REG) break; ++ pfe_eth_enet_addr_byte_mac(ha->addr, &spec_addr); ++ gemac_set_laddrN(priv->EMAC_baseaddr, &spec_addr, uc_count + 2); ++ uc_count++; ++ } ++ } else if ((netdev_mc_count(dev) > 0) || (netdev_uc_count(dev))) { ++ u8 *addr; ++ ++ hash_addr.bottom = 0; ++ hash_addr.top = 0; ++ ++ netdev_for_each_mc_addr(ha, dev) { ++ addr = ha->addr; ++ ++ netif_info(priv, drv, dev, "adding multicast address %X:%X:%X:%X:%X:%X to gem filter\n", ++ addr[0], addr[1], addr[2], addr[3], addr[4], addr[5]); ++ ++ result = pfe_eth_get_hash(addr); ++ ++ if (result >= EMAC_HASH_REG_BITS) { ++ break; ++ } else { ++ if (result < 32) { ++ hash_addr.bottom |= (1 << result); ++ } else { ++ hash_addr.top |= (1 << (result - 32)); ++ } ++ } ++ ++ } ++ ++ uc_count = -1; ++ netdev_for_each_uc_addr(ha, dev) { ++ addr = ha->addr; ++ ++ if(++uc_count < MAX_UC_SPEC_ADDR_REG) ++ { ++ netdev_info(dev, "adding unicast address %02x:%02x:%02x:%02x:%02x:%02x to gem filter\n", ++ addr[0], addr[1], addr[2], ++ addr[3], addr[4], addr[5]); ++ ++ pfe_eth_enet_addr_byte_mac(addr, &spec_addr); ++ gemac_set_laddrN(priv->EMAC_baseaddr, &spec_addr, uc_count + 2); ++ } ++ else ++ { ++ netif_info(priv, drv, dev, "adding unicast address %02x:%02x:%02x:%02x:%02x:%02x to gem hash\n", ++ addr[0], addr[1], addr[2], ++ addr[3], addr[4], addr[5]); ++ ++ result = pfe_eth_get_hash(addr); ++ if (result >= EMAC_HASH_REG_BITS) { ++ break; ++ } else { ++ if (result < 32) ++ hash_addr.bottom |= (1 << result); ++ else ++ hash_addr.top |= (1 << (result - 32)); ++ } ++ ++ ++ } ++ } ++ ++ gemac_set_hash(priv->EMAC_baseaddr, &hash_addr); ++ if(netdev_mc_count(dev)) ++ gemac_enable_multicast(priv->EMAC_baseaddr); ++ else ++ gemac_disable_multicast(priv->EMAC_baseaddr); ++ } ++ ++ if(netdev_uc_count(dev) >= MAX_UC_SPEC_ADDR_REG) ++ gemac_enable_unicast(priv->EMAC_baseaddr); ++ else ++ { ++ /* Check if there are any specific address HW registers that need ++ * to be flushed ++ * */ ++ for(uc_count = netdev_uc_count(dev); uc_count < MAX_UC_SPEC_ADDR_REG; uc_count++) ++ gemac_clear_laddrN(priv->EMAC_baseaddr, uc_count + 2); ++ ++ gemac_disable_unicast(priv->EMAC_baseaddr); ++ } ++ ++ if (dev->flags & IFF_LOOPBACK) { ++ gemac_set_loop(priv->EMAC_baseaddr, LB_LOCAL); ++ } ++ ++ return; ++} ++ ++/** pfe_eth_set_features ++ */ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,6,0) ++static int pfe_eth_set_features(struct net_device *dev, netdev_features_t features) ++#else ++static int pfe_eth_set_features(struct net_device *dev, u32 features) ++#endif ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ int rc = 0; ++ ++ if (features & NETIF_F_RXCSUM) ++ gemac_enable_rx_checksum_offload(priv->EMAC_baseaddr); ++ else ++ gemac_disable_rx_checksum_offload(priv->EMAC_baseaddr); ++ return rc; ++} ++ ++/** pfe_eth_fix_features ++ */ ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,6,0) ++static netdev_features_t pfe_eth_fix_features(struct net_device *dev, netdev_features_t features) ++#else ++static unsigned int pfe_eth_fix_features(struct net_device *dev,u32 features) ++#endif ++{ ++ struct pfe_eth_priv_s *priv = netdev_priv(dev); ++ ++ if (dev->mtu > 1500) ++ { ++ if (features & (NETIF_F_TSO)) ++ { ++ priv->usr_features |= NETIF_F_TSO; ++ features &= ~(NETIF_F_TSO); ++ netdev_err(dev, "TSO cannot be enabled when the MTU is larger than 1500. Please set the MTU to 1500 or lower first.\n"); ++ } ++ } ++ ++ return features; ++} ++ ++/** pfe_eth_tx_timeout ++ */ ++void pfe_eth_tx_timeout(unsigned long data ) ++{ ++ struct pfe_eth_priv_s *priv = (struct pfe_eth_priv_s *)data; ++ ++ netif_info(priv, timer, priv->dev, "%s\n", __func__); ++ ++ pfe_eth_flush_tx(priv, 0); ++ ++ priv->tx_timer.expires = jiffies + ( COMCERTO_TX_RECOVERY_TIMEOUT_MS * HZ )/1000; ++ add_timer(&priv->tx_timer); ++} ++ ++/** pfe_eth_fast_tx_timeout ++ */ ++static enum hrtimer_restart pfe_eth_fast_tx_timeout(struct hrtimer *timer) ++{ ++ struct pfe_eth_fast_timer *fast_tx_timeout = container_of(timer, struct pfe_eth_fast_timer, timer); ++ struct pfe_eth_priv_s *priv = container_of(fast_tx_timeout->base, struct pfe_eth_priv_s, fast_tx_timeout); ++ struct netdev_queue *tx_queue = netdev_get_tx_queue(priv->dev, fast_tx_timeout->queuenum); ++ ++ if(netif_tx_queue_stopped(tx_queue)) { ++#ifdef PFE_ETH_TX_STATS ++ priv->was_stopped[fast_tx_timeout->queuenum] = 1; ++#endif ++ netif_tx_wake_queue(tx_queue); ++ } ++ ++ return HRTIMER_NORESTART; ++} ++ ++/** pfe_eth_fast_tx_timeout_init ++ */ ++static void pfe_eth_fast_tx_timeout_init(struct pfe_eth_priv_s *priv) ++{ ++ int i; ++ for (i = 0; i < emac_txq_cnt; i++) { ++ priv->fast_tx_timeout[i].queuenum = i; ++ hrtimer_init(&priv->fast_tx_timeout[i].timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL); ++ priv->fast_tx_timeout[i].timer.function = pfe_eth_fast_tx_timeout; ++ priv->fast_tx_timeout[i].base = priv->fast_tx_timeout; ++ } ++} ++ ++static struct sk_buff *pfe_eth_rx_skb(struct net_device *dev, struct pfe_eth_priv_s *priv, unsigned int qno) ++{ ++ void *buf_addr; ++ unsigned int rx_ctrl; ++ unsigned int desc_ctrl = 0; ++ struct hif_ipsec_hdr *ipsec_hdr = NULL; ++ struct sk_buff *skb; ++ struct sk_buff *skb_frag, *skb_frag_last = NULL; ++ int length = 0, offset; ++ ++ skb = priv->skb_inflight[qno]; ++ ++ if (skb && (skb_frag_last = skb_shinfo(skb)->frag_list)) { ++ while (skb_frag_last->next) ++ skb_frag_last = skb_frag_last->next; ++ } ++ ++ while (!(desc_ctrl & CL_DESC_LAST)) { ++ ++ buf_addr = hif_lib_receive_pkt(&priv->client, qno, &length, &offset, &rx_ctrl, &desc_ctrl, (void **)&ipsec_hdr); ++ if (!buf_addr) ++ goto incomplete; ++ ++#ifdef PFE_ETH_NAPI_STATS ++ priv->napi_counters[NAPI_DESC_COUNT]++; ++#endif ++ ++ /* First frag */ ++ if (desc_ctrl & CL_DESC_FIRST) { ++#if defined(CONFIG_PLATFORM_EMULATION) || defined(CONFIG_PLATFORM_PCI) ++ skb = dev_alloc_skb(PFE_BUF_SIZE); ++ if (unlikely(!skb)) { ++ goto pkt_drop; ++ } ++ ++ skb_copy_to_linear_data(skb, buf_addr, length + offset); ++ kfree(buf_addr); ++#else ++#if defined(CONFIG_COMCERTO_ZONE_DMA_NCNB) ++ skb = alloc_skb(length + offset + 32, GFP_ATOMIC); ++#else ++ skb = alloc_skb_header(PFE_BUF_SIZE, buf_addr, GFP_ATOMIC); ++#endif ++ if (unlikely(!skb)) { ++ goto pkt_drop; ++ } ++#endif ++ skb_reserve(skb, offset); ++#if defined(CONFIG_COMCERTO_ZONE_DMA_NCNB) ++ __memcpy(skb->data, buf_addr + offset, length); ++ if (ipsec_hdr) { ++ sah_local = *(unsigned int *)&ipsec_hdr->sa_handle[0]; ++ } ++ kfree(buf_addr); ++#endif ++ skb_put(skb, length); ++ skb->dev = dev; ++ ++ if ((dev->features & NETIF_F_RXCSUM) && (rx_ctrl & HIF_CTRL_RX_CHECKSUMMED)) ++ skb->ip_summed = CHECKSUM_UNNECESSARY; ++ else ++ skb_checksum_none_assert(skb); ++ ++ } else { ++ ++ /* Next frags */ ++ if (unlikely(!skb)) { ++ printk(KERN_ERR "%s: NULL skb_inflight\n", __func__); ++ goto pkt_drop; ++ } ++ ++#if defined(CONFIG_COMCERTO_ZONE_DMA_NCNB) ++ skb_frag = alloc_skb(length + offset + 32, GFP_ATOMIC); ++#else ++ skb_frag = alloc_skb_header(PFE_BUF_SIZE, buf_addr, GFP_ATOMIC); ++#endif ++ if (unlikely(!skb_frag)) { ++ kfree(buf_addr); ++ goto pkt_drop; ++ } ++ ++ skb_reserve(skb_frag, offset); ++#if defined(CONFIG_COMCERTO_ZONE_DMA_NCNB) ++ __memcpy(skb_frag->data, buf_addr + offset, length); ++ kfree(buf_addr); ++#endif ++ skb_put(skb_frag, length); ++ ++ skb_frag->dev = dev; ++ ++ if (skb_shinfo(skb)->frag_list) ++ skb_frag_last->next = skb_frag; ++ else ++ skb_shinfo(skb)->frag_list = skb_frag; ++ ++ skb->truesize += skb_frag->truesize; ++ skb->data_len += length; ++ skb->len += length; ++ skb_frag_last = skb_frag; ++ } ++ } ++ ++ priv->skb_inflight[qno] = NULL; ++ return skb; ++ ++incomplete: ++ priv->skb_inflight[qno] = skb; ++ return NULL; ++ ++pkt_drop: ++ priv->skb_inflight[qno] = NULL; ++ ++ if (skb) { ++ kfree_skb(skb); ++ } else { ++ kfree(buf_addr); ++ } ++ ++ priv->stats.rx_errors++; ++ ++ return NULL; ++} ++ ++ ++/** pfe_eth_poll ++ */ ++static int pfe_eth_poll(struct pfe_eth_priv_s *priv, struct napi_struct *napi, unsigned int qno, int budget) ++{ ++ struct net_device *dev = priv->dev; ++ struct sk_buff *skb; ++ int work_done = 0; ++ unsigned int len; ++ ++ netif_info(priv, intr, priv->dev, "%s\n", __func__); ++ ++#ifdef PFE_ETH_NAPI_STATS ++ priv->napi_counters[NAPI_POLL_COUNT]++; ++#endif ++ ++ do { ++ skb = pfe_eth_rx_skb(dev, priv, qno); ++ ++ if (!skb) ++ break; ++ ++ len = skb->len; ++ ++ /* Packet will be processed */ ++ skb->protocol = eth_type_trans(skb, dev); ++ ++ netif_receive_skb(skb); ++ ++ priv->stats.rx_packets++; ++ priv->stats.rx_bytes += len; ++ ++ dev->last_rx = jiffies; ++ ++ work_done++; ++ ++#ifdef PFE_ETH_NAPI_STATS ++ priv->napi_counters[NAPI_PACKET_COUNT]++; ++#endif ++ ++ } while (work_done < budget); ++ ++ /* If no Rx receive nor cleanup work was done, exit polling mode. ++ * No more netif_running(dev) check is required here , as this is checked in ++ * net/core/dev.c ( 2.6.33.5 kernel specific). ++ */ ++ if (work_done < budget) { ++ napi_complete(napi); ++ ++ hif_lib_event_handler_start(&priv->client, EVENT_RX_PKT_IND, qno); ++ } ++#ifdef PFE_ETH_NAPI_STATS ++ else ++ priv->napi_counters[NAPI_FULL_BUDGET_COUNT]++; ++#endif ++ ++ return work_done; ++} ++ ++/** pfe_eth_lro_poll ++ */ ++static int pfe_eth_lro_poll(struct napi_struct *napi, int budget) ++{ ++ struct pfe_eth_priv_s *priv = container_of(napi, struct pfe_eth_priv_s, lro_napi); ++ ++ netif_info(priv, intr, priv->dev, "%s\n", __func__); ++ ++ return pfe_eth_poll(priv, napi, 2, budget); ++} ++ ++ ++/** pfe_eth_low_poll ++ */ ++static int pfe_eth_low_poll(struct napi_struct *napi, int budget) ++{ ++ struct pfe_eth_priv_s *priv = container_of(napi, struct pfe_eth_priv_s, low_napi); ++ ++ netif_info(priv, intr, priv->dev, "%s\n", __func__); ++ ++ return pfe_eth_poll(priv, napi, 1, budget); ++} ++ ++/** pfe_eth_high_poll ++ */ ++static int pfe_eth_high_poll(struct napi_struct *napi, int budget ) ++{ ++ struct pfe_eth_priv_s *priv = container_of(napi, struct pfe_eth_priv_s, high_napi); ++ ++ netif_info(priv, intr, priv->dev, "%s\n", __func__); ++ ++ return pfe_eth_poll(priv, napi, 0, budget); ++} ++ ++static const struct net_device_ops pfe_netdev_ops = { ++ .ndo_open = pfe_eth_open, ++ .ndo_stop = pfe_eth_close, ++ .ndo_start_xmit = pfe_eth_send_packet, ++ .ndo_select_queue = pfe_eth_select_queue, ++ .ndo_get_stats = pfe_eth_get_stats, ++ .ndo_change_mtu = pfe_eth_change_mtu, ++ .ndo_set_mac_address = pfe_eth_set_mac_address, ++ .ndo_set_rx_mode = pfe_eth_set_multi, ++ .ndo_set_features = pfe_eth_set_features, ++ .ndo_fix_features = pfe_eth_fix_features, ++ .ndo_validate_addr = eth_validate_addr, ++}; ++ ++ ++/** pfe_eth_init_one ++ */ ++ ++static int pfe_eth_init_one( struct pfe *pfe, int id ) ++{ ++ struct net_device *dev = NULL; ++ struct pfe_eth_priv_s *priv = NULL; ++ struct comcerto_eth_platform_data *einfo; ++ struct comcerto_mdio_platform_data *minfo; ++ struct comcerto_pfe_platform_data *pfe_info; ++ int err; ++ ++ /* Extract pltform data */ ++#if defined(CONFIG_PLATFORM_EMULATION) || defined(CONFIG_PLATFORM_PCI) ++ pfe_info = (struct comcerto_pfe_platform_data *) &comcerto_pfe_pdata; ++#else ++ pfe_info = (struct comcerto_pfe_platform_data *) pfe->dev->platform_data; ++#endif ++ if (!pfe_info) { ++ printk(KERN_ERR "%s: pfe missing additional platform data\n", __func__); ++ err = -ENODEV; ++ goto err0; ++ } ++ ++ einfo = (struct comcerto_eth_platform_data *) pfe_info->comcerto_eth_pdata; ++ ++ /* einfo never be NULL, but no harm in having this check */ ++ if (!einfo) { ++ printk(KERN_ERR "%s: pfe missing additional gemacs platform data\n", __func__); ++ err = -ENODEV; ++ goto err0; ++ } ++ ++ minfo = (struct comcerto_mdio_platform_data *) pfe_info->comcerto_mdio_pdata; ++ ++ /* einfo never be NULL, but no harm in having this check */ ++ if (!minfo) { ++ printk(KERN_ERR "%s: pfe missing additional mdios platform data\n", __func__); ++ err = -ENODEV; ++ goto err0; ++ } ++ ++ /* ++ * FIXME: Need to check some flag in "einfo" to know whether ++ * GEMAC is enabled Or not. ++ */ ++ ++ /* Create an ethernet device instance */ ++ dev = alloc_etherdev_mq(sizeof (*priv), emac_txq_cnt); ++ ++ if (!dev) { ++ printk(KERN_ERR "%s: gemac %d device allocation failed\n", __func__, einfo[id].gem_id); ++ err = -ENOMEM; ++ goto err0; ++ } ++ ++ priv = netdev_priv(dev); ++ priv->dev = dev; ++ priv->id = einfo[id].gem_id; ++ priv->pfe = pfe; ++ ++#if defined(CONFIG_PLATFORM_C2000) ++ /* get gemac tx clock */ ++ priv->gemtx_clk = clk_get(NULL, "gemtx"); ++ ++ if (IS_ERR(priv->gemtx_clk)) { ++ printk(KERN_ERR "%s: Unable to get the clock for gemac %d\n", __func__, priv->id); ++ err = -ENODEV; ++ goto err1; ++ } ++#endif ++ ++ pfe->eth.eth_priv[id] = priv; ++ ++ /* Set the info in the priv to the current info */ ++ priv->einfo = &einfo[id]; ++ priv->EMAC_baseaddr = cbus_emac_base[id]; ++ priv->PHY_baseaddr = cbus_emac_base[0]; ++ priv->mdio_muxval = einfo[id].mdio_muxval; ++ priv->GPI_baseaddr = cbus_gpi_base[id]; ++ ++ /* FIXME : For now TMU queue numbers hardcoded, later should be taken from pfe.h */ ++#define HIF_GEMAC_TMUQ_BASE 6 ++ priv->low_tmuQ = HIF_GEMAC_TMUQ_BASE + (id * 2); ++ priv->high_tmuQ = priv->low_tmuQ + 1; ++ ++ spin_lock_init(&priv->lock); ++ priv->tx_timer.data = (unsigned long)priv; ++ priv->tx_timer.function = pfe_eth_tx_timeout; ++ priv->tx_timer.expires = jiffies + ( COMCERTO_TX_RECOVERY_TIMEOUT_MS * HZ )/1000; ++ init_timer(&priv->tx_timer); ++ ++ pfe_eth_fast_tx_timeout_init(priv); ++ ++ /* Copy the station address into the dev structure, */ ++ memcpy(dev->dev_addr, einfo[id].mac_addr, ETH_ALEN); ++ ++ /* Initialize mdio */ ++ if (minfo[id].enabled) { ++ if ((err = pfe_eth_mdio_init(priv, &minfo[id]))) { ++ netdev_err(dev, "%s: pfe_eth_mdio_init() failed\n", __func__); ++ goto err2; ++ } ++ } ++ ++ dev->mtu = 1500; ++ ++ /* supported features */ ++ dev->hw_features = NETIF_F_SG; ++ /* Enable after checksum offload is validated ++ dev->hw_features = NETIF_F_RXCSUM | NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM | ++ NETIF_F_SG; */ ++ ++ /* enabled by default */ ++ dev->features = dev->hw_features; ++ ++ priv->usr_features = dev->features; ++ ++ dev->netdev_ops = &pfe_netdev_ops; ++ ++ dev->ethtool_ops = &pfe_ethtool_ops; ++ ++ /* Enable basic messages by default */ ++ priv->msg_enable = NETIF_MSG_IFUP | NETIF_MSG_IFDOWN | NETIF_MSG_LINK | NETIF_MSG_PROBE; ++ ++ netif_napi_add(dev, &priv->low_napi, pfe_eth_low_poll, HIF_RX_POLL_WEIGHT - 16); ++ netif_napi_add(dev, &priv->high_napi, pfe_eth_high_poll, HIF_RX_POLL_WEIGHT - 16); ++ netif_napi_add(dev, &priv->lro_napi, pfe_eth_lro_poll, HIF_RX_POLL_WEIGHT - 16); ++ ++ err = register_netdev(dev); ++ ++ if (err) { ++ netdev_err(dev, "register_netdev() failed\n"); ++ goto err3; ++ } ++ ++ if (!(priv->einfo->phy_flags & GEMAC_NO_PHY)) { ++ err = pfe_phy_init(dev); ++ if (err) { ++ netdev_err(dev, "%s: pfe_phy_init() failed\n", __func__); ++ goto err4; ++ } ++ } ++ ++ ++ /* Create all the sysfs files */ ++ if(pfe_eth_sysfs_init(dev)) ++ goto err4; ++ ++ netif_info(priv, probe, dev, "%s: created interface, baseaddr: %p\n", __func__, priv->EMAC_baseaddr); ++ ++ return 0; ++err4: ++ unregister_netdev(dev); ++err3: ++ pfe_eth_mdio_exit(priv->mii_bus); ++err2: ++#if defined(CONFIG_PLATFORM_C2000) ++ clk_put(priv->gemtx_clk); ++err1: ++#endif ++ free_netdev(priv->dev); ++ ++err0: ++ return err; ++} ++ ++/** pfe_eth_init ++ */ ++int pfe_eth_init(struct pfe *pfe) ++{ ++ int ii = 0; ++ int err; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ cbus_emac_base[0] = EMAC1_BASE_ADDR; ++ cbus_emac_base[1] = EMAC2_BASE_ADDR; ++ ++ cbus_gpi_base[0] = EGPI1_BASE_ADDR; ++ cbus_gpi_base[1] = EGPI2_BASE_ADDR; ++ ++#if !defined(CONFIG_PLATFORM_LS1012A) ++ cbus_emac_base[2] = EMAC3_BASE_ADDR; ++ cbus_gpi_base[2] = EGPI3_BASE_ADDR; ++#endif ++ ++ for (ii = 0; ii < NUM_GEMAC_SUPPORT; ii++) { ++ if ((err = pfe_eth_init_one(pfe, ii))) ++ goto err0; ++ } ++ ++ return 0; ++ ++err0: ++ while(ii--){ ++ pfe_eth_exit_one( pfe->eth.eth_priv[ii] ); ++ } ++ ++ /* Register three network devices in the kernel */ ++ return err; ++} ++ ++/** pfe_eth_exit_one ++ */ ++static void pfe_eth_exit_one(struct pfe_eth_priv_s *priv) ++{ ++ netif_info(priv, probe, priv->dev, "%s\n", __func__); ++ ++ pfe_eth_sysfs_exit(priv->dev); ++ ++#if defined(CONFIG_PLATFORM_C2000) ++ clk_put(priv->gemtx_clk); ++#endif ++ ++ unregister_netdev(priv->dev); ++ ++ pfe_eth_mdio_exit(priv->mii_bus); ++ ++ if (!(priv->einfo->phy_flags & GEMAC_NO_PHY)) ++ pfe_phy_exit(priv->dev); ++ ++ free_netdev(priv->dev); ++} ++ ++/** pfe_eth_exit ++ */ ++void pfe_eth_exit(struct pfe *pfe) ++{ ++ int ii; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ for(ii = 0; ii < NUM_GEMAC_SUPPORT; ii++ ) { ++ /* ++ * FIXME: Need to check some flag in "einfo" to know whether ++ * GEMAC is enabled Or not. ++ */ ++ ++ pfe_eth_exit_one(pfe->eth.eth_priv[ii]); ++ } ++} ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_eth.h +@@ -0,0 +1,384 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_ETH_H_ ++#define _PFE_ETH_H_ ++#include <linux/kernel.h> ++#include <linux/netdevice.h> ++#include <linux/etherdevice.h> ++#include <linux/ethtool.h> ++#include <linux/mii.h> ++#include <linux/phy.h> ++#include <linux/clk.h> ++#include <linux/interrupt.h> ++#include <linux/time.h> ++ ++#define PFE_ETH_NAPI_STATS ++#define PFE_ETH_TX_STATS ++ ++#define PFE_ETH_FRAGS_MAX (65536/HIF_RX_PKT_MIN_SIZE) ++#define LRO_LEN_COUNT_MAX 32 ++#define LRO_NB_COUNT_MAX 32 ++ ++#if defined(CONFIG_PLATFORM_PCI) || defined(CONFIG_PLATFORM_EMULATION) || defined(CONFIG_PLATFORM_LS1012A) ++ ++#define CONFIG_COMCERTO_GEMAC 1 ++ ++#define CONFIG_COMCERTO_USE_MII 1 ++#define CONFIG_COMCERTO_USE_RMII 2 ++#define CONFIG_COMCERTO_USE_GMII 4 ++#define CONFIG_COMCERTO_USE_RGMII 8 ++#define CONFIG_COMCERTO_USE_SGMII 16 ++ ++#define GEMAC_SW_CONF (1 << 8) | (1 << 11) // GEMAC configured by SW ++#define GEMAC_PHY_CONF 0 // GEMAC configured by phy lines (not for MII/GMII) ++#define GEMAC_SW_FULL_DUPLEX (1 << 9) ++#define GEMAC_SW_SPEED_10M (0 << 12) ++#define GEMAC_SW_SPEED_100M (1 << 12) ++#define GEMAC_SW_SPEED_1G (2 << 12) ++ ++#define GEMAC_NO_PHY (1 << 0) // set if no phy connected to MAC (ex ethernet switch). In this case use MAC fixed configuration ++#define GEMAC_PHY_RGMII_ADD_DELAY (1 << 1) ++ ++/* gemac to interface name assignment */ ++#define GEMAC0_ITF_NAME "eth5" ++#define GEMAC1_ITF_NAME "eth6" ++#define GEMAC2_ITF_NAME "eth7" ++ ++#define GEMAC0_MAC { 0x00, 0xED, 0xCD, 0xEF, 0xAA, 0xCC } ++#define GEMAC1_MAC { 0x00, 0x2A, 0x2B, 0x2C, 0x2D, 0x2E } ++ ++struct comcerto_eth_platform_data { ++ /* device specific information */ ++ u32 device_flags; ++ char name[16]; ++ ++ ++ /* board specific information */ ++ u32 mii_config; ++ u32 gemac_mode; ++ u32 phy_flags; ++ u32 gem_id; ++ u32 bus_id; ++ u32 phy_id; ++ u32 mdio_muxval; ++ u8 mac_addr[ETH_ALEN]; ++}; ++ ++struct comcerto_mdio_platform_data { ++ int enabled; ++ int irq[32]; ++ u32 phy_mask; ++ int mdc_div; ++}; ++ ++struct comcerto_pfe_platform_data ++{ ++ struct comcerto_eth_platform_data comcerto_eth_pdata[3]; ++ struct comcerto_mdio_platform_data comcerto_mdio_pdata[3]; ++}; ++#if !defined(CONFIG_PLATFORM_LS1012A) ++static struct comcerto_pfe_platform_data comcerto_pfe_pdata = { ++ .comcerto_eth_pdata[0] = { ++ .name = GEMAC0_ITF_NAME, ++ .device_flags = CONFIG_COMCERTO_GEMAC, ++ .mii_config = CONFIG_COMCERTO_USE_MII, ++ .gemac_mode = GEMAC_SW_CONF | GEMAC_SW_FULL_DUPLEX | GEMAC_SW_SPEED_100M, ++#if defined(CONFIG_PLATFORM_EMULATION) || defined(CONFIG_PLATFORM_PCI) ++ .phy_flags = GEMAC_NO_PHY, ++#else ++ .phy_flags = GEMAC_PHY_RGMII_ADD_DELAY, ++#endif ++ .bus_id = 0, ++ .phy_id = 0, ++ .gem_id = 0, ++ .mac_addr = (u8[])GEMAC0_MAC, ++ }, ++ ++ .comcerto_eth_pdata[1] = { ++ .name = GEMAC1_ITF_NAME, ++ .device_flags = CONFIG_COMCERTO_GEMAC, ++ .mii_config = CONFIG_COMCERTO_USE_RGMII, ++ .gemac_mode = GEMAC_SW_CONF | GEMAC_SW_FULL_DUPLEX | GEMAC_SW_SPEED_1G, ++ .phy_flags = GEMAC_NO_PHY, ++ .gem_id = 1, ++ .mac_addr = (u8[])GEMAC1_MAC, ++ }, ++ ++ .comcerto_eth_pdata[2] = { ++ .name = GEMAC2_ITF_NAME, ++ }, ++ ++ .comcerto_mdio_pdata[0] = { ++ .enabled = 1, ++ .phy_mask = 0xFFFFFFFE, ++ .mdc_div = 96, ++ .irq = { ++ [0] = PHY_POLL, ++ }, ++ }, ++}; ++#endif ++#endif ++ ++#if defined(CONFIG_PLATFORM_LS1012A) ++#define NUM_GEMAC_SUPPORT 2 ++#define DRV_NAME "ls1012a-geth" ++#else ++#define NUM_GEMAC_SUPPORT 3 ++#define DRV_NAME "c2000-geth" ++#endif ++#define COMCERTO_INFOSTR_LEN 32 ++#define COMCERTO_TX_RECOVERY_TIMEOUT_MS 500 ++#define COMCERTO_TX_FAST_RECOVERY_TIMEOUT_MS 3 ++#define TX_POLL_TIMEOUT_MS 1000 ++ ++#define EMAC_TXQ_CNT 16 ++#define EMAC_TXQ_DEPTH (HIF_TX_DESC_NT) ++ ++#define JUMBO_FRAME_SIZE 10258 ++/** ++ * Client Tx queue threshold, for txQ flush condition. ++ * It must be smaller than the queue size (in case we ever change it in the future). ++ */ ++#define HIF_CL_TX_FLUSH_MARK 32 ++ ++/** ++ * Max number of TX resources (HIF descriptors or skbs) that will be released ++ * in a single go during batch recycling. ++ * Should be lower than the flush mark so the SW can provide the HW with a ++ * continuous stream of packets instead of bursts. ++ */ ++#define TX_FREE_MAX_COUNT 16 ++#define EMAC_RXQ_CNT 3 ++#define EMAC_RXQ_DEPTH HIF_RX_DESC_NT /* make sure clients can receive a full burst of packets */ ++#define EMAC_RMON_TXBYTES_POS 0x00 ++#define EMAC_RMON_RXBYTES_POS 0x14 ++ ++#define EMAC_QUEUENUM_MASK (emac_txq_cnt - 1) ++#define EMAC_MDIO_TIMEOUT 1000 ++#define MAX_UC_SPEC_ADDR_REG 31 ++ ++ ++/* The set of statistics registers implemented in the Cadence MAC. ++ * The statistics registers implemented are a subset of all the statistics ++ * available, but contains all the compulsory ones. ++ * For full descriptions on the registers, refer to the Cadence MAC programmers ++ * guide or the IEEE 802.3 specifications. ++ */ ++struct gemac_stats{ ++ u32 octets_tx_bot; /* Lower 32-bits for number of octets tx'd */ ++ u32 octets_tx_top; /* Upper 16-bits for number of octets tx'd */ ++ u32 frames_tx; /* Number of frames transmitted OK */ ++ u32 broadcast_tx; /* Number of broadcast frames transmitted */ ++ u32 multicast_tx; /* Number of multicast frames transmitted */ ++ u32 pause_tx; /* Number of pause frames transmitted. */ ++ u32 frame64_tx; /* Number of 64byte frames transmitted */ ++ u32 frame65_127_tx; /* Number of 65-127 byte frames transmitted */ ++ u32 frame128_255_tx; /* Number of 128-255 byte frames transmitted */ ++ u32 frame256_511_tx; /* Number of 256-511 byte frames transmitted */ ++ u32 frame512_1023_tx; /* Number of 512-1023 byte frames transmitted */ ++ u32 frame1024_1518_tx; /* Number of 1024-1518 byte frames transmitted*/ ++ u32 frame1519_tx; /* Number of frames greater than 1518 bytes tx*/ ++ u32 tx_urun; /* Transmit underrun errors due to DMA */ ++ u32 single_col; /* Number of single collision frames */ ++ u32 multi_col; /* Number of multi collision frames */ ++ u32 excess_col; /* Number of excessive collision frames. */ ++ u32 late_col; /* Collisions occuring after slot time */ ++ u32 def_tx; /* Frames deferred due to crs */ ++ u32 crs_errors; /* Errors caused by crs not being asserted. */ ++ u32 octets_rx_bot; /* Lower 32-bits for number of octets rx'd */ ++ u32 octets_rx_top; /* Upper 16-bits for number of octets rx'd */ ++ u32 frames_rx; /* Number of frames received OK */ ++ u32 broadcast_rx; /* Number of broadcast frames received */ ++ u32 multicast_rx; /* Number of multicast frames received */ ++ u32 pause_rx; /* Number of pause frames received. */ ++ u32 frame64_rx; /* Number of 64byte frames received */ ++ u32 frame65_127_rx; /* Number of 65-127 byte frames received */ ++ u32 frame128_255_rx; /* Number of 128-255 byte frames received */ ++ u32 frame256_511_rx; /* Number of 256-511 byte frames received */ ++ u32 frame512_1023_rx; /* Number of 512-1023 byte frames received */ ++ u32 frame1024_1518_rx; /* Number of 1024-1518 byte frames received*/ ++ u32 frame1519_rx; /* Number of frames greater than 1518 bytes rx*/ ++ u32 usize_frames; /* Frames received less than min of 64 bytes */ ++ u32 excess_length; /* Number of excessive length frames rx */ ++ u32 jabbers; /* Excessive length + crc or align errors. */ ++ u32 fcs_errors; /* Number of frames received with crc errors */ ++ u32 length_check_errors;/* Number of frames with incorrect length */ ++ u32 rx_symbol_errors; /* Number of times rx_er asserted during rx */ ++ u32 align_errors; /* Frames received without integer no. bytes */ ++ u32 rx_res_errors; /* Number of times buffers ran out during rx */ ++ u32 rx_orun; /* Receive overrun errors due to DMA */ ++ u32 ip_cksum; /* IP header checksum errors */ ++ u32 tcp_cksum; /* TCP checksum errors */ ++ u32 udp_cksum; /* UDP checksum errors */ ++}; ++ ++#define EMAC_REG_SPACE sizeof(struct gemac_reg) ++#define EMAC_RMON_LEN (sizeof(struct gemac_stats)/sizeof(u32)) ++ ++ ++struct pfe_eth_fast_timer { ++ int queuenum; ++ struct hrtimer timer; ++ void * base; ++}; ++ ++typedef struct pfe_eth_priv_s ++{ ++ struct pfe *pfe; ++ struct hif_client_s client; ++ struct napi_struct lro_napi; ++ struct napi_struct low_napi; ++ struct napi_struct high_napi; ++ int low_tmuQ; ++ int high_tmuQ; ++ struct net_device_stats stats; ++ struct net_device *dev; ++ int id; ++ int promisc; ++ unsigned int msg_enable; ++ unsigned int usr_features; ++ ++ spinlock_t lock; ++ unsigned int event_status; ++ int irq; ++ void* EMAC_baseaddr; ++ void* PHY_baseaddr; /* This points to the EMAC base from where we access PHY */ ++ void* GPI_baseaddr; ++ int mdio_muxval; ++ /* PHY stuff */ ++ struct phy_device *phydev; ++ int oldspeed; ++ int oldduplex; ++ int oldlink; ++ /* mdio info */ ++ int mdc_div; ++ struct mii_bus *mii_bus; ++ struct clk *gemtx_clk; ++ int wol; ++ ++ int default_priority; ++ struct timer_list tx_timer; ++ struct pfe_eth_fast_timer fast_tx_timeout[EMAC_TXQ_CNT]; ++ ++ struct comcerto_eth_platform_data *einfo; ++ struct sk_buff *skb_inflight[EMAC_RXQ_CNT + 6]; ++ ++#ifdef PFE_ETH_LRO_STATS ++ unsigned int lro_len_counters[LRO_LEN_COUNT_MAX]; ++ unsigned int lro_nb_counters[LRO_NB_COUNT_MAX]; //TODO change to exact max number when RX scatter done ++#endif ++ ++ ++#ifdef PFE_ETH_TX_STATS ++ unsigned int stop_queue_total[EMAC_TXQ_CNT]; ++ unsigned int stop_queue_hif[EMAC_TXQ_CNT]; ++ unsigned int stop_queue_hif_client[EMAC_TXQ_CNT]; ++ unsigned int stop_queue_credit[EMAC_TXQ_CNT]; ++ unsigned int clean_fail[EMAC_TXQ_CNT]; ++ unsigned int was_stopped[EMAC_TXQ_CNT]; ++#endif ++ ++#ifdef PFE_ETH_NAPI_STATS ++ unsigned int napi_counters[NAPI_MAX_COUNT]; ++#endif ++ unsigned int frags_inflight[EMAC_RXQ_CNT + 6]; ++ ++}pfe_eth_priv_t; ++ ++struct pfe_eth { ++ struct pfe_eth_priv_s *eth_priv[3]; ++}; ++ ++int pfe_eth_init(struct pfe *pfe); ++void pfe_eth_exit(struct pfe *pfe); ++int pfe_eth_suspend(struct net_device *dev); ++int pfe_eth_resume(struct net_device *dev); ++int pfe_eth_mdio_reset(struct mii_bus *bus); ++ ++/** pfe_compute_csum ++ * ++ */ ++static int inline pfe_compute_csum(struct sk_buff *skb) ++{ ++ struct skb_shared_info *sh; ++ unsigned int nr_frags; ++ skb_frag_t *f; ++ u32 csum = 0; ++ int i; ++ int len; ++ ++ /* Make sure that no intermediate buffers/fragments are odd byte aligned */ ++ if (skb_is_nonlinear(skb)) { ++ int linearize = 0; ++ ++ sh = skb_shinfo(skb); ++ nr_frags = sh->nr_frags; ++ len = skb_headlen(skb) - skb_transport_offset(skb); ++ ++ if (len & 0x1) { ++ linearize = 1; ++ //printk("#1 Odd length %d\n", len); ++ } ++ else { ++ for (i = 0; i < nr_frags - 1; i++) { ++ f = &sh->frags[i]; ++ len = skb_frag_size(f); ++ ++ if (len & 0x1) { ++ linearize = 1; ++ //printk("#2 %d Odd length %d\n", i, len); ++ break; ++ } ++ } ++ } ++ ++ if (linearize) ++ if (skb_linearize(skb)) ++ return -1; ++ } ++ ++ /* Compute checksum */ ++ if (!skb_is_nonlinear(skb)) { ++ *(u16*)(skb_transport_header(skb) + skb->csum_offset) = csum_fold(csum_partial(skb_transport_header(skb), skb->len - skb_transport_offset(skb), 0)); ++ } ++ else { ++ sh = skb_shinfo(skb); ++ nr_frags = sh->nr_frags; ++ ++ if (nr_frags) { ++ csum = csum_partial(skb_transport_header(skb), skb_headlen(skb) - skb_transport_offset(skb), 0); ++ ++ for (i = 0; i < nr_frags - 1; i++) { ++ f = &sh->frags[i]; ++ csum = csum_partial(skb_frag_address(f), skb_frag_size(f), csum); ++ } ++ ++ f = &sh->frags[i]; ++ *(u16*)(skb_transport_header(skb) + skb->csum_offset) = csum_fold(csum_partial(skb_frag_address(f), skb_frag_size(f), csum)); ++ } ++ } ++ ++ return 0; ++} ++ ++ ++ ++#endif /* _PFE_ETH_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_firmware.c +@@ -0,0 +1,322 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++/** @file ++ * Contains all the functions to handle parsing and loading of PE firmware files. ++ */ ++#include <linux/firmware.h> ++ ++#include "pfe_mod.h" ++#include "pfe_firmware.h" ++#include "pfe/pfe.h" ++ ++static Elf32_Shdr * get_elf_section_header(const struct firmware *fw, const char *section) ++{ ++ Elf32_Ehdr *elf_hdr = (Elf32_Ehdr *)fw->data; ++ Elf32_Shdr *shdr, *shdr_shstr; ++ Elf32_Off e_shoff = be32_to_cpu(elf_hdr->e_shoff); ++ Elf32_Half e_shentsize = be16_to_cpu(elf_hdr->e_shentsize); ++ Elf32_Half e_shnum = be16_to_cpu(elf_hdr->e_shnum); ++ Elf32_Half e_shstrndx = be16_to_cpu(elf_hdr->e_shstrndx); ++ Elf32_Off shstr_offset; ++ Elf32_Word sh_name; ++ const char *name; ++ int i; ++ ++ /* Section header strings */ ++ shdr_shstr = (Elf32_Shdr *)(fw->data + e_shoff + e_shstrndx * e_shentsize); ++ shstr_offset = be32_to_cpu(shdr_shstr->sh_offset); ++ ++ for (i = 0; i < e_shnum; i++) { ++ shdr = (Elf32_Shdr *)(fw->data + e_shoff + i * e_shentsize); ++ ++ sh_name = be32_to_cpu(shdr->sh_name); ++ ++ name = (const char *)(fw->data + shstr_offset + sh_name); ++ ++ if (!strcmp(name, section)) ++ return shdr; ++ } ++ ++ printk(KERN_ERR "%s: didn't find section %s\n", __func__, section); ++ ++ return NULL; ++} ++ ++static unsigned long get_elf_section(const struct firmware *fw, const char *section) ++{ ++ Elf32_Shdr *shdr = get_elf_section_header(fw, section); ++ ++ if (shdr) ++ return be32_to_cpu(shdr->sh_addr); ++ else ++ return -1; ++} ++ ++#if defined(CFG_DIAGS) ++static int pfe_get_diags_info(const struct firmware *fw, struct pfe_diags_info *diags_info) ++{ ++ Elf32_Shdr *shdr; ++ unsigned long offset, size; ++ ++ shdr = get_elf_section_header(fw, ".pfe_diags_str"); ++ if (shdr) ++ { ++ offset = be32_to_cpu(shdr->sh_offset); ++ size = be32_to_cpu(shdr->sh_size); ++ diags_info->diags_str_base = be32_to_cpu(shdr->sh_addr); ++ diags_info->diags_str_size = size; ++ diags_info->diags_str_array = pfe_kmalloc(size, GFP_KERNEL); ++ memcpy(diags_info->diags_str_array, fw->data+offset, size); ++ ++ return 0; ++ } else ++ { ++ return -1; ++ } ++} ++#endif ++ ++static void pfe_check_version_info(const struct firmware *fw) ++{ ++ static char *version = NULL; ++ ++ Elf32_Shdr *shdr = get_elf_section_header(fw, ".version"); ++ ++ if (shdr) ++ { ++ if(!version) ++ { ++ /* this is the first fw we load, use its version string as reference (whatever it is) */ ++ version = (char *)(fw->data + be32_to_cpu(shdr->sh_offset)); ++ ++ printk(KERN_INFO "PFE binary version: %s\n", version); ++ } ++ else ++ { ++ /* already have loaded at least one firmware, check sequence can start now */ ++ if(strcmp(version, (char *)(fw->data + be32_to_cpu(shdr->sh_offset)))) ++ { ++ printk(KERN_INFO "WARNING: PFE firmware binaries from incompatible version\n"); ++ } ++ } ++ } ++ else ++ { ++ /* version cannot be verified, a potential issue that should be reported */ ++ printk(KERN_INFO "WARNING: PFE firmware binaries from incompatible version\n"); ++ } ++} ++ ++/** PFE elf firmware loader. ++* Loads an elf firmware image into a list of PE's (specified using a bitmask) ++* ++* @param pe_mask Mask of PE id's to load firmware to ++* @param fw Pointer to the firmware image ++* ++* @return 0 on sucess, a negative value on error ++* ++*/ ++int pfe_load_elf(int pe_mask, const struct firmware *fw, struct pfe *pfe) ++{ ++ Elf32_Ehdr *elf_hdr = (Elf32_Ehdr *)fw->data; ++ Elf32_Half sections = be16_to_cpu(elf_hdr->e_shnum); ++ Elf32_Shdr *shdr = (Elf32_Shdr *) (fw->data + be32_to_cpu(elf_hdr->e_shoff)); ++ int id, section; ++ int rc; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ /* Some sanity checks */ ++ if (strncmp(&elf_hdr->e_ident[EI_MAG0], ELFMAG, SELFMAG)) ++ { ++ printk(KERN_ERR "%s: incorrect elf magic number\n", __func__); ++ return -EINVAL; ++ } ++ ++ if (elf_hdr->e_ident[EI_CLASS] != ELFCLASS32) ++ { ++ printk(KERN_ERR "%s: incorrect elf class(%x)\n", __func__, elf_hdr->e_ident[EI_CLASS]); ++ return -EINVAL; ++ } ++ ++ if (elf_hdr->e_ident[EI_DATA] != ELFDATA2MSB) ++ { ++ printk(KERN_ERR "%s: incorrect elf data(%x)\n", __func__, elf_hdr->e_ident[EI_DATA]); ++ return -EINVAL; ++ } ++ ++ if (be16_to_cpu(elf_hdr->e_type) != ET_EXEC) ++ { ++ printk(KERN_ERR "%s: incorrect elf file type(%x)\n", __func__, be16_to_cpu(elf_hdr->e_type)); ++ return -EINVAL; ++ } ++ ++ for (section = 0; section < sections; section++, shdr++) ++ { ++ if (!(be32_to_cpu(shdr->sh_flags) & (SHF_WRITE | SHF_ALLOC | SHF_EXECINSTR))) ++ continue; ++ ++ for (id = 0; id < MAX_PE; id++) ++ if (pe_mask & (1 << id)) ++ { ++ rc = pe_load_elf_section(id, fw->data, shdr, pfe->dev); ++ if (rc < 0) ++ goto err; ++ } ++ } ++ ++ pfe_check_version_info(fw); ++ ++ return 0; ++ ++err: ++ return rc; ++} ++ ++ ++/** PFE firmware initialization. ++* Loads different firmware files from filesystem. ++* Initializes PE IMEM/DMEM and UTIL-PE DDR ++* Initializes control path symbol addresses (by looking them up in the elf firmware files ++* Takes PE's out of reset ++* ++* @return 0 on sucess, a negative value on error ++* ++*/ ++int pfe_firmware_init(struct pfe *pfe) ++{ ++ const struct firmware *class_fw, *tmu_fw; ++ int rc = 0; ++#if !defined(CONFIG_UTIL_DISABLED) ++ const char* util_fw_name; ++ const struct firmware *util_fw; ++#endif ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ if (request_firmware(&class_fw, CLASS_FIRMWARE_FILENAME, pfe->dev)) { ++ printk(KERN_ERR "%s: request firmware %s failed\n", __func__, CLASS_FIRMWARE_FILENAME); ++ rc = -ETIMEDOUT; ++ goto err0; ++ } ++ ++ if (request_firmware(&tmu_fw, TMU_FIRMWARE_FILENAME, pfe->dev)) { ++ printk(KERN_ERR "%s: request firmware %s failed\n", __func__, TMU_FIRMWARE_FILENAME); ++ rc = -ETIMEDOUT; ++ goto err1; ++ } ++#if !defined(CONFIG_UTIL_DISABLED) ++#if defined(CONFIG_PLATFORM_C2000) ++ util_fw_name = (system_rev == 0) ? UTIL_REVA0_FIRMWARE_FILENAME : UTIL_FIRMWARE_FILENAME; ++#else ++ util_fw_name = UTIL_FIRMWARE_FILENAME; ++#endif ++ ++ if (request_firmware(&util_fw, util_fw_name, pfe->dev)) { ++ printk(KERN_ERR "%s: request firmware %s failed\n", __func__, util_fw_name); ++ rc = -ETIMEDOUT; ++ goto err2; ++ } ++#endif ++ rc = pfe_load_elf(CLASS_MASK, class_fw, pfe); ++ if (rc < 0) { ++ printk(KERN_ERR "%s: class firmware load failed\n", __func__); ++ goto err3; ++ } ++ ++ pfe->ctrl.class_dmem_sh = get_elf_section(class_fw, ".dmem_sh"); ++ pfe->ctrl.class_pe_lmem_sh = get_elf_section(class_fw, ".pe_lmem_sh"); ++ ++#if defined(CFG_DIAGS) ++ rc = pfe_get_diags_info(class_fw, &pfe->diags.class_diags_info); ++ if (rc < 0) { ++ printk (KERN_WARNING "PFE diags won't be available for class PEs\n"); ++ rc = 0; ++ } ++#endif ++ ++ printk(KERN_INFO "%s: class firmware loaded %#lx %#lx\n", __func__, pfe->ctrl.class_dmem_sh, pfe->ctrl.class_pe_lmem_sh); ++ ++ rc = pfe_load_elf(TMU_MASK, tmu_fw, pfe); ++ if (rc < 0) { ++ printk(KERN_ERR "%s: tmu firmware load failed\n", __func__); ++ goto err3; ++ } ++ ++ pfe->ctrl.tmu_dmem_sh = get_elf_section(tmu_fw, ".dmem_sh"); ++ ++ printk(KERN_INFO "%s: tmu firmware loaded %#lx\n", __func__, pfe->ctrl.tmu_dmem_sh); ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ rc = pfe_load_elf(UTIL_MASK, util_fw, pfe); ++ if (rc < 0) { ++ printk(KERN_ERR "%s: util firmware load failed\n", __func__); ++ goto err3; ++ } ++ ++ pfe->ctrl.util_dmem_sh = get_elf_section(util_fw, ".dmem_sh"); ++ pfe->ctrl.util_ddr_sh = get_elf_section(util_fw, ".ddr_sh"); ++ ++#if defined(CFG_DIAGS) ++ rc = pfe_get_diags_info(util_fw, &pfe->diags.util_diags_info); ++ if (rc < 0) { ++ printk(KERN_WARNING "PFE diags won't be available for util PE\n"); ++ rc = 0; ++ } ++#endif ++ ++ printk(KERN_INFO "%s: util firmware loaded %#lx\n", __func__, pfe->ctrl.util_dmem_sh); ++ ++ util_enable(); ++#endif ++ ++ tmu_enable(0xf); ++ class_enable(); ++ ++err3: ++#if !defined(CONFIG_UTIL_DISABLED) ++ release_firmware(util_fw); ++ ++err2: ++#endif ++ release_firmware(tmu_fw); ++ ++err1: ++ release_firmware(class_fw); ++ ++err0: ++ return rc; ++} ++ ++/** PFE firmware cleanup ++* Puts PE's in reset ++* ++* ++*/ ++void pfe_firmware_exit(struct pfe *pfe) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ class_disable(); ++ tmu_disable(0xf); ++#if !defined(CONFIG_UTIL_DISABLED) ++ util_disable(); ++#endif ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_firmware.h +@@ -0,0 +1,41 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_FIRMWARE_H_ ++#define _PFE_FIRMWARE_H_ ++ ++#if defined(CONFIG_PLATFORM_C2000) ++#define CLASS_FIRMWARE_FILENAME "class_c2000.elf" ++#define TMU_FIRMWARE_FILENAME "tmu_c2000.elf" ++#define UTIL_FIRMWARE_FILENAME "util_c2000.elf" ++#define UTIL_REVA0_FIRMWARE_FILENAME "util_c2000_revA0.elf" ++#else ++#define CLASS_FIRMWARE_FILENAME "ppfe_class_ls1012a.elf" ++#define TMU_FIRMWARE_FILENAME "ppfe_tmu_ls1012a.elf" ++#endif ++ ++#define PFE_FW_CHECK_PASS 0 ++#define PFE_FW_CHECK_FAIL 1 ++#define NUM_PFE_FW 3 ++ ++int pfe_firmware_init(struct pfe *pfe); ++void pfe_firmware_exit(struct pfe *pfe); ++ ++#endif /* _PFE_FIRMWARE_H_ */ ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hal.c +@@ -0,0 +1,2217 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++ ++#include "pfe_ctrl_hal.h" ++#include "pfe/pfe.h" ++ ++void *cbus_base_addr; ++void *ddr_base_addr; ++unsigned long ddr_phys_base_addr; ++unsigned int ddr_size; ++ ++static struct pe_info pe[MAX_PE]; ++ ++/** Initializes the PFE library. ++* Must be called before using any of the library functions. ++* ++* @param[in] cbus_base CBUS virtual base address (as mapped in the host CPU address space) ++* @param[in] ddr_base PFE DDR range virtual base address (as mapped in the host CPU address space) ++* @param[in] ddr_phys_base PFE DDR range physical base address (as mapped in platform) ++* @param[in] size PFE DDR range size (as defined by the host software) ++*/ ++void pfe_lib_init(void *cbus_base, void *ddr_base, unsigned long ddr_phys_base, unsigned int size) ++{ ++ cbus_base_addr = cbus_base; ++ ddr_base_addr = ddr_base; ++ ddr_phys_base_addr = ddr_phys_base; ++ ddr_size = size; ++ ++ pe[CLASS0_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(0); ++ pe[CLASS0_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(0); ++ pe[CLASS0_ID].pmem_size = CLASS_IMEM_SIZE; ++ pe[CLASS0_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA; ++ pe[CLASS0_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR; ++ pe[CLASS0_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA; ++ ++ pe[CLASS1_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(1); ++ pe[CLASS1_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(1); ++ pe[CLASS1_ID].pmem_size = CLASS_IMEM_SIZE; ++ pe[CLASS1_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA; ++ pe[CLASS1_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR; ++ pe[CLASS1_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA; ++ ++ pe[CLASS2_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(2); ++ pe[CLASS2_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(2); ++ pe[CLASS2_ID].pmem_size = CLASS_IMEM_SIZE; ++ pe[CLASS2_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA; ++ pe[CLASS2_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR; ++ pe[CLASS2_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA; ++ ++ pe[CLASS3_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(3); ++ pe[CLASS3_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(3); ++ pe[CLASS3_ID].pmem_size = CLASS_IMEM_SIZE; ++ pe[CLASS3_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA; ++ pe[CLASS3_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR; ++ pe[CLASS3_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA; ++ ++#if !defined(CONFIG_PLATFORM_PCI) ++ pe[CLASS4_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(4); ++ pe[CLASS4_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(4); ++ pe[CLASS4_ID].pmem_size = CLASS_IMEM_SIZE; ++ pe[CLASS4_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA; ++ pe[CLASS4_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR; ++ pe[CLASS4_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA; ++ ++ pe[CLASS5_ID].dmem_base_addr = CLASS_DMEM_BASE_ADDR(5); ++ pe[CLASS5_ID].pmem_base_addr = CLASS_IMEM_BASE_ADDR(5); ++ pe[CLASS5_ID].pmem_size = CLASS_IMEM_SIZE; ++ pe[CLASS5_ID].mem_access_wdata = CLASS_MEM_ACCESS_WDATA; ++ pe[CLASS5_ID].mem_access_addr = CLASS_MEM_ACCESS_ADDR; ++ pe[CLASS5_ID].mem_access_rdata = CLASS_MEM_ACCESS_RDATA; ++#endif ++ pe[TMU0_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(0); ++ pe[TMU0_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(0); ++ pe[TMU0_ID].pmem_size = TMU_IMEM_SIZE; ++ pe[TMU0_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA; ++ pe[TMU0_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR; ++ pe[TMU0_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA; ++ ++#if !defined(CONFIG_TMU_DUMMY) ++ pe[TMU1_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(1); ++ pe[TMU1_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(1); ++ pe[TMU1_ID].pmem_size = TMU_IMEM_SIZE; ++ pe[TMU1_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA; ++ pe[TMU1_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR; ++ pe[TMU1_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA; ++ ++#if !defined(CONFIG_PLATFORM_LS1012A) ++ pe[TMU2_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(2); ++ pe[TMU2_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(2); ++ pe[TMU2_ID].pmem_size = TMU_IMEM_SIZE; ++ pe[TMU2_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA; ++ pe[TMU2_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR; ++ pe[TMU2_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA; ++#endif ++ ++ pe[TMU3_ID].dmem_base_addr = TMU_DMEM_BASE_ADDR(3); ++ pe[TMU3_ID].pmem_base_addr = TMU_IMEM_BASE_ADDR(3); ++ pe[TMU3_ID].pmem_size = TMU_IMEM_SIZE; ++ pe[TMU3_ID].mem_access_wdata = TMU_MEM_ACCESS_WDATA; ++ pe[TMU3_ID].mem_access_addr = TMU_MEM_ACCESS_ADDR; ++ pe[TMU3_ID].mem_access_rdata = TMU_MEM_ACCESS_RDATA; ++#endif ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ pe[UTIL_ID].dmem_base_addr = UTIL_DMEM_BASE_ADDR; ++ pe[UTIL_ID].mem_access_wdata = UTIL_MEM_ACCESS_WDATA; ++ pe[UTIL_ID].mem_access_addr = UTIL_MEM_ACCESS_ADDR; ++ pe[UTIL_ID].mem_access_rdata = UTIL_MEM_ACCESS_RDATA; ++#endif ++} ++ ++ ++/** Writes a buffer to PE internal memory from the host ++ * through indirect access registers. ++ * ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] src Buffer source address ++ * @param[in] mem_access_addr DMEM destination address (must be 32bit aligned) ++ * @param[in] len Number of bytes to copy ++ */ ++void pe_mem_memcpy_to32(int id, u32 mem_access_addr, const void *src, unsigned int len) ++{ ++ u32 offset = 0, val, addr; ++ unsigned int len32 = len >> 2; ++ int i; ++ ++ addr = mem_access_addr | PE_MEM_ACCESS_WRITE | PE_MEM_ACCESS_BYTE_ENABLE(0, 4); ++ ++ for (i = 0; i < len32; i++, offset += 4, src += 4) { ++ val = *(u32 *)src; ++ writel(cpu_to_be32(val), pe[id].mem_access_wdata); ++ writel(addr + offset, pe[id].mem_access_addr); ++ } ++ ++ if ((len = (len & 0x3))) { ++ val = 0; ++ ++ addr = (mem_access_addr | PE_MEM_ACCESS_WRITE | PE_MEM_ACCESS_BYTE_ENABLE(0, len)) + offset; ++ ++ for (i = 0; i < len; i++, src++) ++ val |= (*(u8 *)src) << (8 * i); ++ ++ writel(cpu_to_be32(val), pe[id].mem_access_wdata); ++ writel(addr, pe[id].mem_access_addr); ++ } ++} ++ ++/** Writes a buffer to PE internal data memory (DMEM) from the host ++ * through indirect access registers. ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] src Buffer source address ++ * @param[in] dst DMEM destination address (must be 32bit aligned) ++ * @param[in] len Number of bytes to copy ++ */ ++void pe_dmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len) ++{ ++ pe_mem_memcpy_to32(id, pe[id].dmem_base_addr | dst | PE_MEM_ACCESS_DMEM, src, len); ++} ++ ++ ++/** Writes a buffer to PE internal program memory (PMEM) from the host ++ * through indirect access registers. ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., TMU3_ID) ++ * @param[in] src Buffer source address ++ * @param[in] dst PMEM destination address (must be 32bit aligned) ++ * @param[in] len Number of bytes to copy ++ */ ++void pe_pmem_memcpy_to32(int id, u32 dst, const void *src, unsigned int len) ++{ ++ pe_mem_memcpy_to32(id, pe[id].pmem_base_addr | (dst & (pe[id].pmem_size - 1)) | PE_MEM_ACCESS_IMEM, src, len); ++} ++ ++ ++/** Reads PE internal program memory (IMEM) from the host ++ * through indirect access registers. ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., TMU3_ID) ++ * @param[in] addr PMEM read address (must be aligned on size) ++ * @param[in] size Number of bytes to read (maximum 4, must not cross 32bit boundaries) ++ * @return the data read (in PE endianess, i.e BE). ++ */ ++u32 pe_pmem_read(int id, u32 addr, u8 size) ++{ ++ u32 offset = addr & 0x3; ++ u32 mask = 0xffffffff >> ((4 - size) << 3); ++ u32 val; ++ ++ addr = pe[id].pmem_base_addr | ((addr & ~0x3) & (pe[id].pmem_size - 1)) | PE_MEM_ACCESS_IMEM | PE_MEM_ACCESS_BYTE_ENABLE(offset, size); ++ ++ writel(addr, pe[id].mem_access_addr); ++ val = be32_to_cpu(readl(pe[id].mem_access_rdata)); ++ ++ return (val >> (offset << 3)) & mask; ++} ++ ++ ++/** Writes PE internal data memory (DMEM) from the host ++ * through indirect access registers. ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] addr DMEM write address (must be aligned on size) ++ * @param[in] val Value to write (in PE endianess, i.e BE) ++ * @param[in] size Number of bytes to write (maximum 4, must not cross 32bit boundaries) ++ */ ++void pe_dmem_write(int id, u32 val, u32 addr, u8 size) ++{ ++ u32 offset = addr & 0x3; ++ ++ addr = pe[id].dmem_base_addr | (addr & ~0x3) | PE_MEM_ACCESS_WRITE | PE_MEM_ACCESS_DMEM | PE_MEM_ACCESS_BYTE_ENABLE(offset, size); ++ ++ /* Indirect access interface is byte swapping data being written */ ++ writel(cpu_to_be32(val << (offset << 3)), pe[id].mem_access_wdata); ++ writel(addr, pe[id].mem_access_addr); ++} ++ ++ ++/** Reads PE internal data memory (DMEM) from the host ++ * through indirect access registers. ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] addr DMEM read address (must be aligned on size) ++ * @param[in] size Number of bytes to read (maximum 4, must not cross 32bit boundaries) ++ * @return the data read (in PE endianess, i.e BE). ++ */ ++u32 pe_dmem_read(int id, u32 addr, u8 size) ++{ ++ u32 offset = addr & 0x3; ++ u32 mask = 0xffffffff >> ((4 - size) << 3); ++ u32 val; ++ ++ addr = pe[id].dmem_base_addr | (addr & ~0x3) | PE_MEM_ACCESS_DMEM | PE_MEM_ACCESS_BYTE_ENABLE(offset, size); ++ ++ writel(addr, pe[id].mem_access_addr); ++ ++ /* Indirect access interface is byte swapping data being read */ ++ val = be32_to_cpu(readl(pe[id].mem_access_rdata)); ++ ++ return (val >> (offset << 3)) & mask; ++} ++ ++ ++/** This function is used to write to CLASS internal bus peripherals (ccu, pe-lem) from the host ++* through indirect access registers. ++* @param[in] val value to write ++* @param[in] addr Address to write to (must be aligned on size) ++* @param[in] size Number of bytes to write (1, 2 or 4) ++* ++*/ ++void class_bus_write(u32 val, u32 addr, u8 size) ++{ ++ u32 offset = addr & 0x3; ++ ++ writel((addr & CLASS_BUS_ACCESS_BASE_MASK), CLASS_BUS_ACCESS_BASE); ++ ++ addr = (addr & ~CLASS_BUS_ACCESS_BASE_MASK) | PE_MEM_ACCESS_WRITE | (size << 24); ++ ++ writel(cpu_to_be32(val << (offset << 3)), CLASS_BUS_ACCESS_WDATA); ++ writel(addr, CLASS_BUS_ACCESS_ADDR); ++} ++ ++ ++/** Reads from CLASS internal bus peripherals (ccu, pe-lem) from the host ++* through indirect access registers. ++* @param[in] addr Address to read from (must be aligned on size) ++* @param[in] size Number of bytes to read (1, 2 or 4) ++* @return the read data ++* ++*/ ++u32 class_bus_read(u32 addr, u8 size) ++{ ++ u32 offset = addr & 0x3; ++ u32 mask = 0xffffffff >> ((4 - size) << 3); ++ u32 val; ++ ++ writel((addr & CLASS_BUS_ACCESS_BASE_MASK), CLASS_BUS_ACCESS_BASE); ++ ++ addr = (addr & ~CLASS_BUS_ACCESS_BASE_MASK) | (size << 24); ++ ++ writel(addr, CLASS_BUS_ACCESS_ADDR); ++ val = be32_to_cpu(readl(CLASS_BUS_ACCESS_RDATA)); ++ ++ return (val >> (offset << 3)) & mask; ++} ++ ++ ++/** Writes data to the cluster memory (PE_LMEM) ++* @param[in] dst PE LMEM destination address (must be 32bit aligned) ++* @param[in] src Buffer source address ++* @param[in] len Number of bytes to copy ++*/ ++void class_pe_lmem_memcpy_to32(u32 dst, const void *src, unsigned int len) ++{ ++ u32 len32 = len >> 2; ++ int i; ++ ++ for (i = 0; i < len32; i++, src += 4, dst += 4) ++ class_bus_write(*(u32 *)src, dst, 4); ++ ++ if (len & 0x2) ++ { ++ class_bus_write(*(u16 *)src, dst, 2); ++ src += 2; ++ dst += 2; ++ } ++ ++ if (len & 0x1) ++ { ++ class_bus_write(*(u8 *)src, dst, 1); ++ src++; ++ dst++; ++ } ++} ++ ++/** Writes value to the cluster memory (PE_LMEM) ++* @param[in] dst PE LMEM destination address (must be 32bit aligned) ++* @param[in] val Value to write ++* @param[in] len Number of bytes to write ++*/ ++void class_pe_lmem_memset(u32 dst, int val, unsigned int len) ++{ ++ u32 len32 = len >> 2; ++ int i; ++ ++ val = val | (val << 8) | (val << 16) | (val << 24); ++ ++ for (i = 0; i < len32; i++, dst += 4) ++ class_bus_write(val, dst, 4); ++ ++ if (len & 0x2) ++ { ++ class_bus_write(val, dst, 2); ++ dst += 2; ++ } ++ ++ if (len & 0x1) ++ { ++ class_bus_write(val, dst, 1); ++ dst++; ++ } ++} ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ ++/** Writes UTIL program memory (DDR) from the host. ++ * ++ * @param[in] addr Address to write (virtual, must be aligned on size) ++ * @param[in] val Value to write (in PE endianess, i.e BE) ++ * @param[in] size Number of bytes to write (2 or 4) ++ */ ++static void util_pmem_write(u32 val, void *addr, u8 size) ++{ ++ void *addr64 = (void *)((unsigned long)addr & ~0x7); ++ unsigned long off = 8 - ((unsigned long)addr & 0x7) - size; ++ ++ //IMEM should be loaded as a 64bit swapped value in a 64bit aligned location ++ if (size == 4) ++ writel(be32_to_cpu(val), addr64 + off); ++ else ++ writew(be16_to_cpu((u16)val), addr64 + off); ++} ++ ++ ++/** Writes a buffer to UTIL program memory (DDR) from the host. ++ * ++ * @param[in] dst Address to write (virtual, must be at least 16bit aligned) ++ * @param[in] src Buffer to write (in PE endianess, i.e BE, must have same alignment as dst) ++ * @param[in] len Number of bytes to write (must be at least 16bit aligned) ++ */ ++static void util_pmem_memcpy(void *dst, const void *src, unsigned int len) ++{ ++ unsigned int len32; ++ int i; ++ ++ if ((unsigned long)src & 0x2) { ++ util_pmem_write(*(u16 *)src, dst, 2); ++ src += 2; ++ dst += 2; ++ len -= 2; ++ } ++ ++ len32 = len >> 2; ++ ++ for (i = 0; i < len32; i++, dst += 4, src += 4) ++ util_pmem_write(*(u32 *)src, dst, 4); ++ ++ if (len & 0x2) ++ util_pmem_write(*(u16 *)src, dst, len & 0x2); ++} ++#endif ++ ++/** Loads an elf section into pmem ++ * Code needs to be at least 16bit aligned and only PROGBITS sections are supported ++ * ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., TMU3_ID) ++ * @param[in] data pointer to the elf firmware ++ * @param[in] shdr pointer to the elf section header ++ * ++ */ ++static int pe_load_pmem_section(int id, const void *data, Elf32_Shdr *shdr) ++{ ++ u32 offset = be32_to_cpu(shdr->sh_offset); ++ u32 addr = be32_to_cpu(shdr->sh_addr); ++ u32 size = be32_to_cpu(shdr->sh_size); ++ u32 type = be32_to_cpu(shdr->sh_type); ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ if (id == UTIL_ID) ++ { ++ printk(KERN_ERR "%s: unsuported pmem section for UTIL\n", __func__); ++ return -EINVAL; ++ } ++#endif ++ ++ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) ++ { ++ printk(KERN_ERR "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n", ++ __func__, addr, (unsigned long) data + offset); ++ ++ return -EINVAL; ++ } ++ ++ if (addr & 0x1) ++ { ++ printk(KERN_ERR "%s: load address(%x) is not 16bit aligned\n", __func__, addr); ++ return -EINVAL; ++ } ++ ++ if (size & 0x1) ++ { ++ printk(KERN_ERR "%s: load size(%x) is not 16bit aligned\n", __func__, size); ++ return -EINVAL; ++ } ++ ++ switch (type) ++ { ++ case SHT_PROGBITS: ++ pe_pmem_memcpy_to32(id, addr, data + offset, size); ++ ++ break; ++ ++ default: ++ printk(KERN_ERR "%s: unsuported section type(%x)\n", __func__, type); ++ return -EINVAL; ++ break; ++ } ++ ++ return 0; ++} ++ ++ ++/** Loads an elf section into dmem ++ * Data needs to be at least 32bit aligned, NOBITS sections are correctly initialized to 0 ++ * ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] data pointer to the elf firmware ++ * @param[in] shdr pointer to the elf section header ++ * ++ */ ++static int pe_load_dmem_section(int id, const void *data, Elf32_Shdr *shdr) ++{ ++ u32 offset = be32_to_cpu(shdr->sh_offset); ++ u32 addr = be32_to_cpu(shdr->sh_addr); ++ u32 size = be32_to_cpu(shdr->sh_size); ++ u32 type = be32_to_cpu(shdr->sh_type); ++ u32 size32 = size >> 2; ++ int i; ++ ++ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) ++ { ++ printk(KERN_ERR "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n", ++ __func__, addr, (unsigned long)data + offset); ++ ++ return -EINVAL; ++ } ++ ++ if (addr & 0x3) ++ { ++ printk(KERN_ERR "%s: load address(%x) is not 32bit aligned\n", __func__, addr); ++ return -EINVAL; ++ } ++ ++ switch (type) ++ { ++ case SHT_PROGBITS: ++ pe_dmem_memcpy_to32(id, addr, data + offset, size); ++ break; ++ ++ case SHT_NOBITS: ++ for (i = 0; i < size32; i++, addr += 4) ++ pe_dmem_write(id, 0, addr, 4); ++ ++ if (size & 0x3) ++ pe_dmem_write(id, 0, addr, size & 0x3); ++ ++ break; ++ ++ default: ++ printk(KERN_ERR "%s: unsuported section type(%x)\n", __func__, type); ++ return -EINVAL; ++ break; ++ } ++ ++ return 0; ++} ++ ++ ++/** Loads an elf section into DDR ++ * Data needs to be at least 32bit aligned, NOBITS sections are correctly initialized to 0 ++ * ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] data pointer to the elf firmware ++ * @param[in] shdr pointer to the elf section header ++ * ++ */ ++static int pe_load_ddr_section(int id, const void *data, Elf32_Shdr *shdr, struct device *dev) ++{ ++ u32 offset = be32_to_cpu(shdr->sh_offset); ++ u32 addr = be32_to_cpu(shdr->sh_addr); ++ u32 size = be32_to_cpu(shdr->sh_size); ++ u32 type = be32_to_cpu(shdr->sh_type); ++ u32 flags = be32_to_cpu(shdr->sh_flags); ++ ++ switch (type) ++ { ++ case SHT_PROGBITS: ++ if (flags & SHF_EXECINSTR) ++ { ++ if (id <= CLASS_MAX_ID) ++ { ++ /* DO the loading only once in DDR */ ++ if (id == CLASS0_ID) ++ { ++ printk(KERN_ERR "%s: load address(%x) and elf file address(%lx) rcvd\n", __func__, addr, (unsigned long)data + offset); ++ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) ++ { ++ printk(KERN_ERR "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n", ++ __func__, addr, (unsigned long)data + offset); ++ ++ return -EINVAL; ++ } ++ ++ if (addr & 0x1) ++ { ++ printk(KERN_ERR "%s: load address(%x) is not 16bit aligned\n", __func__, addr); ++ return -EINVAL; ++ } ++ ++ if (size & 0x1) ++ { ++ printk(KERN_ERR "%s: load length(%x) is not 16bit aligned\n", __func__, size); ++ return -EINVAL; ++ } ++ memcpy(DDR_PHYS_TO_VIRT(DDR_PFE_TO_PHYS(addr)), data + offset, size); ++ } ++ } ++#if !defined(CONFIG_UTIL_DISABLED) ++ else if (id == UTIL_ID) ++ { ++ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) ++ { ++ printk(KERN_ERR "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n", ++ __func__, addr, (unsigned long)data + offset); ++ ++ return -EINVAL; ++ } ++ ++ if (addr & 0x1) ++ { ++ printk(KERN_ERR "%s: load address(%x) is not 16bit aligned\n", __func__, addr); ++ return -EINVAL; ++ } ++ ++ if (size & 0x1) ++ { ++ printk(KERN_ERR "%s: load length(%x) is not 16bit aligned\n", __func__, size); ++ return -EINVAL; ++ } ++ ++ util_pmem_memcpy(DDR_PHYS_TO_VIRT(DDR_PFE_TO_PHYS(addr)), data + offset, size); ++ } ++#endif ++ else ++ { ++ printk(KERN_ERR "%s: unsuported ddr section type(%x) for PE(%d)\n", __func__, type, id); ++ return -EINVAL; ++ } ++ ++ } ++ else ++ { ++ memcpy(DDR_PHYS_TO_VIRT(DDR_PFE_TO_PHYS(addr)), data + offset, size); ++ } ++ ++ break; ++ ++ case SHT_NOBITS: ++ memset(DDR_PHYS_TO_VIRT(DDR_PFE_TO_PHYS(addr)), 0, size); ++ ++ break; ++ ++ default: ++ printk(KERN_ERR "%s: unsuported section type(%x)\n", __func__, type); ++ return -EINVAL; ++ break; ++ } ++ ++ return 0; ++} ++ ++ ++/** Loads an elf section into pe lmem ++ * Data needs to be at least 32bit aligned, NOBITS sections are correctly initialized to 0 ++ * ++ * @param[in] id PE identification (CLASS0_ID,..., CLASS5_ID) ++ * @param[in] data pointer to the elf firmware ++ * @param[in] shdr pointer to the elf section header ++ * ++ */ ++static int pe_load_pe_lmem_section(int id, const void *data, Elf32_Shdr *shdr) ++{ ++ u32 offset = be32_to_cpu(shdr->sh_offset); ++ u32 addr = be32_to_cpu(shdr->sh_addr); ++ u32 size = be32_to_cpu(shdr->sh_size); ++ u32 type = be32_to_cpu(shdr->sh_type); ++ ++ if (id > CLASS_MAX_ID) ++ { ++ printk(KERN_ERR "%s: unsuported pe-lmem section type(%x) for PE(%d)\n", __func__, type, id); ++ return -EINVAL; ++ } ++ ++ if (((unsigned long)(data + offset) & 0x3) != (addr & 0x3)) ++ { ++ printk(KERN_ERR "%s: load address(%x) and elf file address(%lx) don't have the same alignment\n", ++ __func__, addr, (unsigned long)data + offset); ++ ++ return -EINVAL; ++ } ++ ++ if (addr & 0x3) ++ { ++ printk(KERN_ERR "%s: load address(%x) is not 32bit aligned\n", __func__, addr); ++ return -EINVAL; ++ } ++ ++ switch (type) ++ { ++ case SHT_PROGBITS: ++ class_pe_lmem_memcpy_to32(addr, data + offset, size); ++ break; ++ ++ case SHT_NOBITS: ++ class_pe_lmem_memset(addr, 0, size); ++ break; ++ ++ default: ++ printk(KERN_ERR "%s: unsuported section type(%x)\n", __func__, type); ++ return -EINVAL; ++ break; ++ } ++ ++ return 0; ++} ++ ++ ++/** Loads an elf section into a PE ++ * For now only supports loading a section to dmem (all PE's), pmem (class and tmu PE's), ++ * DDDR (util PE code) ++ * ++ * @param[in] id PE identification (CLASS0_ID, ..., TMU0_ID, ..., UTIL_ID) ++ * @param[in] data pointer to the elf firmware ++ * @param[in] shdr pointer to the elf section header ++ * ++ */ ++int pe_load_elf_section(int id, const void *data, Elf32_Shdr *shdr, struct device *dev) ++{ ++ u32 addr = be32_to_cpu(shdr->sh_addr); ++ u32 size = be32_to_cpu(shdr->sh_size); ++ ++ if (IS_DMEM(addr, size)) ++ return pe_load_dmem_section(id, data, shdr); ++ else if (IS_PMEM(addr, size)) ++ return pe_load_pmem_section(id, data, shdr); ++ else if (IS_PFE_LMEM(addr, size)) ++ return 0; /* FIXME */ ++ else if (IS_PHYS_DDR(addr, size)) ++ return pe_load_ddr_section(id, data, shdr, dev); ++ else if (IS_PE_LMEM(addr, size)) ++ return pe_load_pe_lmem_section(id, data, shdr); ++ else { ++ printk(KERN_ERR "%s: unsuported memory range(%x)\n", __func__, addr); ++// return -EINVAL; ++ } ++ ++ return 0; ++} ++ ++ ++/**************************** BMU ***************************/ ++ ++/** Initializes a BMU block. ++* @param[in] base BMU block base address ++* @param[in] cfg BMU configuration ++*/ ++void bmu_init(void *base, BMU_CFG *cfg) ++{ ++ bmu_disable(base); ++ ++ bmu_set_config(base, cfg); ++ ++ bmu_reset(base); ++} ++ ++/** Resets a BMU block. ++* @param[in] base BMU block base address ++*/ ++void bmu_reset(void *base) ++{ ++ writel(CORE_SW_RESET, base + BMU_CTRL); ++ ++ /* Wait for self clear */ ++ while (readl(base + BMU_CTRL) & CORE_SW_RESET) ; ++} ++ ++/** Enabled a BMU block. ++* @param[in] base BMU block base address ++*/ ++void bmu_enable(void *base) ++{ ++ writel (CORE_ENABLE, base + BMU_CTRL); ++} ++ ++/** Disables a BMU block. ++* @param[in] base BMU block base address ++*/ ++void bmu_disable(void *base) ++{ ++ writel (CORE_DISABLE, base + BMU_CTRL); ++} ++ ++/** Sets the configuration of a BMU block. ++* @param[in] base BMU block base address ++* @param[in] cfg BMU configuration ++*/ ++void bmu_set_config(void *base, BMU_CFG *cfg) ++{ ++ writel (cfg->baseaddr, base + BMU_UCAST_BASE_ADDR); ++ writel (cfg->count & 0xffff, base + BMU_UCAST_CONFIG); ++ writel (cfg->size & 0xffff, base + BMU_BUF_SIZE); ++// writel (BMU1_THRES_CNT, base + BMU_THRES); ++ ++ /* Interrupts are never used */ ++// writel (0x0, base + BMU_INT_SRC); ++ writel (0x0, base + BMU_INT_ENABLE); ++} ++#if defined(CONFIG_PLATFORM_C2000) ++/**************************** GEMAC ***************************/ ++ ++/** Enable Rx Checksum Engine. With this enabled, Frame with bad IP, ++ * TCP or UDP checksums are discarded ++ * ++ * @param[in] base GEMAC base address. ++ */ ++void gemac_enable_rx_checksum_offload(void *base) ++{ ++ writel(readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_CHKSUM_RX, base + EMAC_NETWORK_CONFIG); ++ writel(readl(CLASS_L4_CHKSUM_ADDR) | IPV4_CHKSUM_DROP, CLASS_L4_CHKSUM_ADDR); ++} ++ ++/** Disable Rx Checksum Engine. ++ * ++ * @param[in] base GEMAC base address. ++ */ ++void gemac_disable_rx_checksum_offload(void *base) ++{ ++ writel(readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_CHKSUM_RX, base + EMAC_NETWORK_CONFIG); ++ writel(readl(CLASS_L4_CHKSUM_ADDR) & ~IPV4_CHKSUM_DROP, CLASS_L4_CHKSUM_ADDR); ++} ++ ++/** Setup the MII Mgmt clock speed. ++ * @param[in] base GEMAC base address (GEMAC0, GEMAC1, GEMAC2) ++ * @param[in] mdc_div MII clock dividor ++ */ ++void gemac_set_mdc_div(void *base, int mdc_div) ++{ ++ u32 val = readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_MDC_DIV_MASK; ++ u32 div; ++ ++ switch (mdc_div) { ++ case 8: ++ div = 0; ++ break; ++ ++ case 16: ++ div = 1; ++ break; ++ ++ case 32: ++ div = 2; ++ break; ++ ++ case 48: ++ div = 3; ++ break; ++ ++ default: ++ case 64: ++ div = 4; ++ break; ++ ++ case 96: ++ div = 5; ++ break; ++ ++ case 128: ++ div = 6; ++ break; ++ ++ case 224: ++ div = 7; ++ break; ++ } ++ ++ val |= div << 18; ++ ++ writel(val, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC set speed. ++* @param[in] base GEMAC base address ++* @param[in] speed GEMAC speed (10, 100 or 1000 Mbps) ++*/ ++void gemac_set_speed(void *base, MAC_SPEED gem_speed) ++{ ++ u32 val = readl(base + EMAC_NETWORK_CONFIG); ++ ++ val = val & ~EMAC_SPEED_MASK; ++ ++ switch (gem_speed) ++ { ++ case SPEED_10M: ++ val &= (~EMAC_PCS_ENABLE); ++ break; ++ ++ case SPEED_100M: ++ val = val | EMAC_SPEED_100; ++ val &= (~EMAC_PCS_ENABLE); ++ break; ++ ++ case SPEED_1000M: ++ val = val | EMAC_SPEED_1000; ++ val &= (~EMAC_PCS_ENABLE); ++ break; ++ ++ case SPEED_1000M_PCS: ++ val = val | EMAC_SPEED_1000; ++ val |= EMAC_PCS_ENABLE; ++ break; ++ ++ default: ++ val = val | EMAC_SPEED_100; ++ val &= (~EMAC_PCS_ENABLE); ++ break; ++ } ++ ++ writel (val, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC set duplex. ++* @param[in] base GEMAC base address ++* @param[in] duplex GEMAC duplex mode (Full, Half) ++*/ ++void gemac_set_duplex(void *base, int duplex) ++{ ++ u32 val = readl(base + EMAC_NETWORK_CONFIG); ++ ++ if (duplex == DUPLEX_HALF) ++ val = (val & ~EMAC_DUPLEX_MASK) | EMAC_HALF_DUP; ++ else ++ val = (val & ~EMAC_DUPLEX_MASK) | EMAC_FULL_DUP; ++ ++ writel (val, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC set mode. ++* @param[in] base GEMAC base address ++* @param[in] mode GEMAC operation mode (MII, RMII, RGMII, SGMII) ++*/ ++ ++#if defined(CONFIG_IP_ALIGNED) ++#define IP_ALIGNED_BITVAL EMAC_TWO_BYTES_IP_ALIGN ++#else ++#define IP_ALIGNED_BITVAL 0 ++#endif ++ ++void gemac_set_mode(void *base, int mode) ++{ ++ switch (mode) ++ { ++ case GMII: ++ writel ((readl(base + EMAC_CONTROL) & ~EMAC_MODE_MASK) | EMAC_GMII_MODE_ENABLE | IP_ALIGNED_BITVAL, base + EMAC_CONTROL); ++ writel (readl(base + EMAC_NETWORK_CONFIG) & (~EMAC_SGMII_MODE_ENABLE), base + EMAC_NETWORK_CONFIG); ++ break; ++ ++ case RGMII: ++ writel ((readl(base + EMAC_CONTROL) & ~EMAC_MODE_MASK) | EMAC_RGMII_MODE_ENABLE | IP_ALIGNED_BITVAL, base + EMAC_CONTROL); ++ writel (readl(base + EMAC_NETWORK_CONFIG) & (~EMAC_SGMII_MODE_ENABLE), base + EMAC_NETWORK_CONFIG); ++ break; ++ ++ case RMII: ++ writel ((readl(base + EMAC_CONTROL) & ~EMAC_MODE_MASK) | EMAC_RMII_MODE_ENABLE | IP_ALIGNED_BITVAL, base + EMAC_CONTROL); ++ writel (readl(base + EMAC_NETWORK_CONFIG) & (~EMAC_SGMII_MODE_ENABLE), base + EMAC_NETWORK_CONFIG); ++ break; ++ ++ case MII: ++ writel ((readl(base + EMAC_CONTROL) & ~EMAC_MODE_MASK) | EMAC_MII_MODE_ENABLE | IP_ALIGNED_BITVAL, base + EMAC_CONTROL); ++ writel (readl(base + EMAC_NETWORK_CONFIG) & (~EMAC_SGMII_MODE_ENABLE), base + EMAC_NETWORK_CONFIG); ++ break; ++ ++ case SGMII: ++ writel ((readl(base + EMAC_CONTROL) & ~EMAC_MODE_MASK) | (EMAC_RMII_MODE_DISABLE | EMAC_RGMII_MODE_DISABLE) | IP_ALIGNED_BITVAL, base + EMAC_CONTROL); ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_SGMII_MODE_ENABLE, base + EMAC_NETWORK_CONFIG); ++ break; ++ ++ default: ++ writel ((readl(base + EMAC_CONTROL) & ~EMAC_MODE_MASK) | EMAC_MII_MODE_ENABLE | IP_ALIGNED_BITVAL, base + EMAC_CONTROL); ++ writel (readl(base + EMAC_NETWORK_CONFIG) & (~EMAC_SGMII_MODE_ENABLE), base + EMAC_NETWORK_CONFIG); ++ break; ++ } ++} ++/** GEMAC Enable MDIO: Activate the Management interface. This is required to program the PHY ++ * @param[in] base GEMAC base address ++ */ ++void gemac_enable_mdio(void *base) ++{ ++ u32 data; ++ ++ data = readl(base + EMAC_NETWORK_CONTROL); ++ data |= EMAC_MDIO_EN; ++ writel(data, base + EMAC_NETWORK_CONTROL); ++} ++ ++/** GEMAC Disable MDIO: Disable the Management interface. ++ * @param[in] base GEMAC base address ++ */ ++void gemac_disable_mdio(void *base) ++{ ++ u32 data; ++ ++ data = readl(base + EMAC_NETWORK_CONTROL); ++ data &= ~EMAC_MDIO_EN; ++ writel(data, base + EMAC_NETWORK_CONTROL); ++} ++ ++ ++/** GEMAC reset function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_reset(void *base) ++{ ++} ++ ++/** GEMAC enable function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONTROL) | EMAC_TX_ENABLE | EMAC_RX_ENABLE, base + EMAC_NETWORK_CONTROL); ++} ++ ++/** GEMAC disable function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONTROL) & ~(EMAC_TX_ENABLE | EMAC_RX_ENABLE), base + EMAC_NETWORK_CONTROL); ++} ++ ++/** GEMAC TX disable function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_tx_disable(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONTROL) & ~(EMAC_TX_ENABLE), base + EMAC_NETWORK_CONTROL); ++} ++ ++/** GEMAC set mac address configuration. ++* @param[in] base GEMAC base address ++* @param[in] addr MAC address to be configured ++*/ ++void gemac_set_address(void *base, SPEC_ADDR *addr) ++{ ++ writel(addr->one.bottom, base + EMAC_SPEC1_ADD_BOT); ++ writel(addr->one.top, base + EMAC_SPEC1_ADD_TOP); ++ writel(addr->two.bottom, base + EMAC_SPEC2_ADD_BOT); ++ writel(addr->two.top, base + EMAC_SPEC2_ADD_TOP); ++ writel(addr->three.bottom, base + EMAC_SPEC3_ADD_BOT); ++ writel(addr->three.top, base + EMAC_SPEC3_ADD_TOP); ++ writel(addr->four.bottom, base + EMAC_SPEC4_ADD_BOT); ++ writel(addr->four.top, base + EMAC_SPEC4_ADD_TOP); ++} ++ ++/** GEMAC get mac address configuration. ++* @param[in] base GEMAC base address ++* ++* @return MAC addresses configured ++*/ ++SPEC_ADDR gemac_get_address(void *base) ++{ ++ SPEC_ADDR addr; ++ ++ addr.one.bottom = readl(base + EMAC_SPEC1_ADD_BOT); ++ addr.one.top = readl(base + EMAC_SPEC1_ADD_TOP); ++ addr.two.bottom = readl(base + EMAC_SPEC2_ADD_BOT); ++ addr.two.top = readl(base + EMAC_SPEC2_ADD_TOP); ++ addr.three.bottom = readl(base + EMAC_SPEC3_ADD_BOT); ++ addr.three.top = readl(base + EMAC_SPEC3_ADD_TOP); ++ addr.four.bottom = readl(base + EMAC_SPEC4_ADD_BOT); ++ addr.four.top = readl(base + EMAC_SPEC4_ADD_TOP); ++ ++ return addr; ++} ++ ++/** Sets the hash register of the MAC. ++ * This register is used for matching unicast and multicast frames. ++ * ++ * @param[in] base GEMAC base address. ++ * @param[in] hash 64-bit hash to be configured. ++ */ ++void gemac_set_hash( void *base, MAC_ADDR *hash ) ++{ ++ writel(hash->bottom, base + EMAC_HASH_BOT); ++ writel(hash->top, base + EMAC_HASH_TOP); ++} ++ ++/** Get the current value hash register of the MAC. ++ * This register is used for matching unicast and multicast frames. ++ * ++ * @param[in] base GEMAC base address ++ ++ * @returns 64-bit hash. ++ */ ++MAC_ADDR gemac_get_hash( void *base ) ++{ ++ MAC_ADDR hash; ++ ++ hash.bottom = readl(base + EMAC_HASH_BOT); ++ hash.top = readl(base + EMAC_HASH_TOP); ++ ++ return hash; ++} ++ ++/** GEMAC set specific local addresses of the MAC. ++* Rather than setting up all four specific addresses, this function sets them up individually. ++* ++* @param[in] base GEMAC base address ++* @param[in] addr MAC address to be configured ++*/ ++void gemac_set_laddr1(void *base, MAC_ADDR *address) ++{ ++ writel(address->bottom, base + EMAC_SPEC1_ADD_BOT); ++ writel(address->top, base + EMAC_SPEC1_ADD_TOP); ++} ++ ++ ++void gemac_set_laddr2(void *base, MAC_ADDR *address) ++{ ++ writel(address->bottom, base + EMAC_SPEC2_ADD_BOT); ++ writel(address->top, base + EMAC_SPEC2_ADD_TOP); ++} ++ ++ ++void gemac_set_laddr3(void *base, MAC_ADDR *address) ++{ ++ writel(address->bottom, base + EMAC_SPEC3_ADD_BOT); ++ writel(address->top, base + EMAC_SPEC3_ADD_TOP); ++} ++ ++ ++void gemac_set_laddr4(void *base, MAC_ADDR *address) ++{ ++ writel(address->bottom, base + EMAC_SPEC4_ADD_BOT); ++ writel(address->top, base + EMAC_SPEC4_ADD_TOP); ++} ++ ++void gemac_set_laddrN(void *base, MAC_ADDR *address, unsigned int entry_index) ++{ ++ if( (entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX) ) ++ return; ++ ++ entry_index = entry_index - 1; ++ ++ if (entry_index < 4) ++ { ++ writel(address->bottom, base + (entry_index * 8) + EMAC_SPEC1_ADD_BOT); ++ writel(address->top, base + (entry_index * 8) + EMAC_SPEC1_ADD_TOP); ++ } ++ else ++ { ++ writel(address->bottom, base + ((entry_index - 4) * 8) + EMAC_SPEC5_ADD_BOT); ++ writel(address->top, base + ((entry_index - 4) * 8) + EMAC_SPEC5_ADD_TOP); ++ } ++} ++ ++/** Get specific local addresses of the MAC. ++* This allows returning of a single specific address stored in the MAC. ++* @param[in] base GEMAC base address ++* ++* @return Specific MAC address 1 ++* ++*/ ++MAC_ADDR gem_get_laddr1(void *base) ++{ ++ MAC_ADDR addr; ++ addr.bottom = readl(base + EMAC_SPEC1_ADD_BOT); ++ addr.top = readl(base + EMAC_SPEC1_ADD_TOP); ++ return addr; ++} ++ ++ ++MAC_ADDR gem_get_laddr2(void *base) ++{ ++ MAC_ADDR addr; ++ addr.bottom = readl(base + EMAC_SPEC2_ADD_BOT); ++ addr.top = readl(base + EMAC_SPEC2_ADD_TOP); ++ return addr; ++} ++ ++ ++MAC_ADDR gem_get_laddr3(void *base) ++{ ++ MAC_ADDR addr; ++ addr.bottom = readl(base + EMAC_SPEC3_ADD_BOT); ++ addr.top = readl(base + EMAC_SPEC3_ADD_TOP); ++ return addr; ++} ++ ++ ++MAC_ADDR gem_get_laddr4(void *base) ++{ ++ MAC_ADDR addr; ++ addr.bottom = readl(base + EMAC_SPEC4_ADD_BOT); ++ addr.top = readl(base + EMAC_SPEC4_ADD_TOP); ++ return addr; ++} ++ ++MAC_ADDR gem_get_laddrN(void *base, unsigned int entry_index) ++{ ++ MAC_ADDR addr = {0xffffffff, 0xffffffff}; ++ ++ if( (entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX) ) ++ return addr; ++ ++ entry_index = entry_index - 1; ++ ++ if (entry_index < 4) ++ { ++ addr.bottom = readl(base + (entry_index * 8) + EMAC_SPEC1_ADD_BOT); ++ addr.top = readl(base + (entry_index * 8) + EMAC_SPEC1_ADD_TOP); ++ } ++ else ++ { ++ addr.bottom = readl(base + ((entry_index - 4) * 8) + EMAC_SPEC5_ADD_BOT); ++ addr.top = readl(base + ((entry_index - 4) * 8) + EMAC_SPEC5_ADD_TOP); ++ } ++ ++ return addr; ++} ++ ++/** Clear specific local addresses of the MAC. ++ * @param[in] base GEMAC base address ++ */ ++ ++void gemac_clear_laddr1(void *base) ++{ ++ writel(0, base + EMAC_SPEC1_ADD_BOT); ++} ++ ++void gemac_clear_laddr2(void *base) ++{ ++ writel(0, base + EMAC_SPEC2_ADD_BOT); ++} ++ ++void gemac_clear_laddr3(void *base) ++{ ++ writel(0, base + EMAC_SPEC3_ADD_BOT); ++} ++ ++void gemac_clear_laddr4(void *base) ++{ ++ writel(0, base + EMAC_SPEC4_ADD_BOT); ++} ++ ++void gemac_clear_laddrN(void *base, unsigned int entry_index) ++{ ++ if( (entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX) ) ++ return; ++ ++ entry_index = entry_index - 1; ++ ++ if ( entry_index < 4 ) ++ writel(0, base + (entry_index * 8) + EMAC_SPEC1_ADD_BOT); ++ else ++ writel(0, base + ((entry_index - 4) * 8) + EMAC_SPEC5_ADD_BOT); ++} ++ ++/** Set the loopback mode of the MAC. This can be either no loopback for normal ++ * operation, local loopback through MAC internal loopback module or PHY ++ * loopback for external loopback through a PHY. This asserts the external loop ++ * pin. ++ * ++ * @param[in] base GEMAC base address. ++ * @param[in] gem_loop Loopback mode to be enabled. LB_LOCAL - MAC Loopback, ++ * LB_EXT - PHY Loopback. ++ */ ++void gemac_set_loop( void *base, MAC_LOOP gem_loop ) ++{ ++ switch (gem_loop) { ++ case LB_LOCAL: ++ writel(readl(base + EMAC_NETWORK_CONTROL) & (~EMAC_LB_PHY), ++ base + EMAC_NETWORK_CONTROL); ++ writel(readl(base + EMAC_NETWORK_CONTROL) | (EMAC_LB_MAC), ++ base + EMAC_NETWORK_CONTROL); ++ break; ++ case LB_EXT: ++ writel(readl(base + EMAC_NETWORK_CONTROL) & (~EMAC_LB_MAC), ++ base + EMAC_NETWORK_CONTROL); ++ writel(readl(base + EMAC_NETWORK_CONTROL) | (EMAC_LB_PHY), ++ base + EMAC_NETWORK_CONTROL); ++ break; ++ default: ++ writel(readl(base + EMAC_NETWORK_CONTROL) & (~(EMAC_LB_MAC | EMAC_LB_PHY)), ++ base + EMAC_NETWORK_CONTROL); ++ } ++} ++ ++/** GEMAC allow frames ++ * @param[in] base GEMAC base address ++ */ ++void gemac_enable_copy_all(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_COPY_ALL, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC do not allow frames ++ * @param[in] base GEMAC base address ++*/ ++void gemac_disable_copy_all(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_COPY_ALL, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC allow broadcast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_allow_broadcast(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_NO_BROADCAST, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC no broadcast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_no_broadcast(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_NO_BROADCAST, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC enable unicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_unicast(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_UNICAST, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC disable unicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_unicast(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_UNICAST, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC enable multicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_multicast(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_MULTICAST, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC disable multicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_multicast(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_MULTICAST, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC enable fcs rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_fcs_rx(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_FCS_RX, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC disable fcs rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_fcs_rx(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_FCS_RX, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC enable 1536 rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_1536_rx(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_1536_RX, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC disable 1536 rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_1536_rx(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_1536_RX, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC enable jumbo function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_rx_jmb(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_JUMBO_FRAME, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC disable jumbo function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_rx_jmb(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_JUMBO_FRAME, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC enable stacked vlan function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_stacked_vlan(void *base) ++{ ++ writel (readl(base + EMAC_STACKED_VLAN_REG) | EMAC_ENABLE_STACKED_VLAN, base + EMAC_STACKED_VLAN_REG); ++} ++ ++/** GEMAC enable stacked vlan function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_stacked_vlan(void *base) ++{ ++ writel (readl(base + EMAC_STACKED_VLAN_REG) & ~EMAC_ENABLE_STACKED_VLAN, base + EMAC_STACKED_VLAN_REG); ++} ++ ++/** GEMAC enable pause rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_pause_rx(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) | EMAC_ENABLE_PAUSE_RX, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC disable pause rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_pause_rx(void *base) ++{ ++ writel (readl(base + EMAC_NETWORK_CONFIG) & ~EMAC_ENABLE_PAUSE_RX, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** GEMAC wol configuration ++* @param[in] base GEMAC base address ++* @param[in] wol_conf WoL register configuration ++*/ ++void gemac_set_wol(void *base, u32 wol_conf) ++{ ++ writel(wol_conf, base + EMAC_WOL); ++} ++ ++/** Sets Gemac bus width to 64bit ++ * @param[in] base GEMAC base address ++ * @param[in] width gemac bus width to be set possible values are 32/64/128 ++ * */ ++void gemac_set_bus_width(void *base, int width) ++{ ++ u32 val = readl(base + EMAC_NETWORK_CONFIG); ++ switch(width) ++ { ++ case 32: ++ val = (val & ~EMAC_DATA_BUS_WIDTH_MASK) | EMAC_DATA_BUS_WIDTH_32; ++ case 128: ++ val = (val & ~EMAC_DATA_BUS_WIDTH_MASK) | EMAC_DATA_BUS_WIDTH_128; ++ case 64: ++ default: ++ val = (val & ~EMAC_DATA_BUS_WIDTH_MASK) | EMAC_DATA_BUS_WIDTH_64; ++ ++ } ++ writel (val, base + EMAC_NETWORK_CONFIG); ++} ++ ++/** Sets Gemac configuration. ++* @param[in] base GEMAC base address ++* @param[in] cfg GEMAC configuration ++*/ ++void gemac_set_config(void *base, GEMAC_CFG *cfg) ++{ ++ gemac_set_mode(base, cfg->mode); ++ ++ gemac_set_speed(base, cfg->speed); ++ ++ gemac_set_duplex(base,cfg->duplex); ++} ++#elif defined(CONFIG_PLATFORM_LS1012A) ++/**************************** MTIP GEMAC ***************************/ ++ ++/** Enable Rx Checksum Engine. With this enabled, Frame with bad IP, ++ * TCP or UDP checksums are discarded ++ * ++ * @param[in] base GEMAC base address. ++ */ ++void gemac_enable_rx_checksum_offload(void *base) ++{ ++ /*Do not find configuration to do this */ ++} ++ ++/** Disable Rx Checksum Engine. ++ * ++ * @param[in] base GEMAC base address. ++ */ ++void gemac_disable_rx_checksum_offload(void *base) ++{ ++ /*Do not find configuration to do this */ ++} ++ ++/** GEMAC set speed. ++* @param[in] base GEMAC base address ++* @param[in] speed GEMAC speed (10, 100 or 1000 Mbps) ++*/ ++void gemac_set_speed(void *base, MAC_SPEED gem_speed) ++{ ++ u32 ecr = readl(base + EMAC_ECNTRL_REG) & ~EMAC_ECNTRL_SPEED; ++ u32 rcr = readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_RMII_10T; ++ ++ switch (gem_speed) ++ { ++ case SPEED_10M: ++ rcr |= EMAC_RCNTRL_RMII_10T; ++ break; ++ ++ ++ case SPEED_1000M: ++ ecr |= EMAC_ECNTRL_SPEED; ++ break; ++ ++ case SPEED_100M: ++ default: ++ /*It is in 100M mode */ ++ break; ++ } ++ writel(ecr, (base + EMAC_ECNTRL_REG)); ++ writel(rcr, (base + EMAC_RCNTRL_REG)); ++} ++ ++/** GEMAC set duplex. ++* @param[in] base GEMAC base address ++* @param[in] duplex GEMAC duplex mode (Full, Half) ++*/ ++void gemac_set_duplex(void *base, int duplex) ++{ ++ ++ if (duplex == DUPLEX_HALF) { ++ printk("%s() TODO\n", __func__); ++ writel(readl(base + EMAC_TCNTRL_REG) & ~EMAC_TCNTRL_FDEN, base + EMAC_TCNTRL_REG); ++ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_DRT, (base + EMAC_RCNTRL_REG)); ++ }else{ ++ writel(readl(base + EMAC_TCNTRL_REG) | EMAC_TCNTRL_FDEN, base + EMAC_TCNTRL_REG); ++ writel(readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_DRT, (base + EMAC_RCNTRL_REG)); ++ } ++} ++ ++/** GEMAC set mode. ++* @param[in] base GEMAC base address ++* @param[in] mode GEMAC operation mode (MII, RMII, RGMII, SGMII) ++*/ ++void gemac_set_mode(void *base, int mode) ++{ ++ u32 val = readl(base + EMAC_RCNTRL_REG); ++ ++ /*Remove loopbank*/ ++ val &= ~EMAC_RCNTRL_LOOP; ++ ++ /*Enable flow control and MII mode*/ ++ val |= (EMAC_RCNTRL_FCE | EMAC_RCNTRL_MII_MODE); ++ ++ writel(val, base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC enable function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable(void *base) ++{ ++ writel(readl(base + EMAC_ECNTRL_REG) | EMAC_ECNTRL_ETHER_EN, base + EMAC_ECNTRL_REG); ++} ++ ++/** GEMAC disable function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable(void *base) ++{ ++ writel(readl(base + EMAC_ECNTRL_REG) & ~EMAC_ECNTRL_ETHER_EN, base + EMAC_ECNTRL_REG); ++} ++ ++/** GEMAC TX disable function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_tx_disable(void *base) ++{ ++ writel(readl(base + EMAC_TCNTRL_REG) | EMAC_TCNTRL_GTS, base + EMAC_TCNTRL_REG); ++} ++ ++/** Sets the hash register of the MAC. ++ * This register is used for matching unicast and multicast frames. ++ * ++ * @param[in] base GEMAC base address. ++ * @param[in] hash 64-bit hash to be configured. ++ */ ++void gemac_set_hash( void *base, MAC_ADDR *hash ) ++{ ++ writel(hash->bottom, base + EMAC_GALR); ++ writel(hash->top, base + EMAC_GAUR); ++} ++ ++void gemac_set_laddrN(void *base, MAC_ADDR *address, unsigned int entry_index) ++{ ++ if( (entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX) ) ++ return; ++ ++ entry_index = entry_index - 1; ++ if (entry_index < 1) { ++ writel(htonl(address->bottom), base + EMAC_PHY_ADDR_LOW); ++ writel((htonl(address->top) | 0x8808), base + EMAC_PHY_ADDR_HIGH); ++ } ++ else ++ { ++ /* TODO for other entry_index */ ++ /*printk("%s for entry_index %d \n",__func__, entry_index); */ ++ writel(htonl(address->bottom), base + ((entry_index - 1) * 8) + EMAC_SMAC_0_0); ++ writel((htonl(address->top) | 0x8808), base + ((entry_index - 1) * 8) + EMAC_SMAC_0_1); ++ } ++ ++} ++ ++void gemac_clear_laddrN(void *base, unsigned int entry_index) ++{ ++ if( (entry_index < 1) || (entry_index > EMAC_SPEC_ADDR_MAX) ) ++ return; ++ ++ entry_index = entry_index - 1; ++ if (entry_index < 1) { ++ writel(0, base + EMAC_PHY_ADDR_LOW); ++ writel(0, base + EMAC_PHY_ADDR_HIGH); ++ } ++ else ++ { ++ writel(0, base + ((entry_index - 1) * 8) + EMAC_SMAC_0_0); ++ writel(0, base + ((entry_index - 1) * 8) + EMAC_SMAC_0_1); ++ } ++ ++ ++} ++ ++/** Set the loopback mode of the MAC. This can be either no loopback for normal ++ * operation, local loopback through MAC internal loopback module or PHY ++ * loopback for external loopback through a PHY. This asserts the external loop ++ * pin. ++ * ++ * @param[in] base GEMAC base address. ++ * @param[in] gem_loop Loopback mode to be enabled. LB_LOCAL - MAC Loopback, ++ * LB_EXT - PHY Loopback. ++ */ ++void gemac_set_loop( void *base, MAC_LOOP gem_loop ) ++{ ++ printk("%s()\n", __func__); ++ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_LOOP, (base + EMAC_RCNTRL_REG)); ++} ++ ++ ++/** GEMAC allow frames ++ * @param[in] base GEMAC base address ++ */ ++void gemac_enable_copy_all(void *base) ++{ ++ writel(readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_PROM, (base + EMAC_RCNTRL_REG)); ++} ++ ++/** GEMAC do not allow frames ++ * @param[in] base GEMAC base address ++*/ ++void gemac_disable_copy_all(void *base) ++{ ++ writel(readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_PROM, (base + EMAC_RCNTRL_REG)); ++} ++ ++/** GEMAC allow broadcast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_allow_broadcast(void *base) ++{ ++ writel (readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_BC_REJ, base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC no broadcast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_no_broadcast(void *base) ++{ ++ writel (readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_BC_REJ, base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC enable unicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_unicast(void *base) ++{ ++ return; ++} ++ ++/** GEMAC disable unicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_unicast(void *base) ++{ ++ return; ++} ++ ++/** GEMAC enable multicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_multicast(void *base) ++{ ++ return; ++} ++ ++/** GEMAC disable multicast function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_multicast(void *base) ++{ ++ /* TODO how to disable multicast? */ ++ return; ++} ++ ++/** GEMAC enable fcs rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_fcs_rx(void *base) ++{ ++ /*Do not find configuration to do this */ ++} ++ ++/** GEMAC disable fcs rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_fcs_rx(void *base) ++{ ++ /*Do not find configuration to do this */ ++} ++ ++ ++/** GEMAC enable 1536 rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_1536_rx(void *base) ++{ ++ /* Set 1536 as Maximum frame length */ ++ writel (readl(base + EMAC_RCNTRL_REG) | (1536 << 16), base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC enable jumbo function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_rx_jmb(void *base) ++{ ++ /*TODO what is the jumbo size supported by MTIP */ ++ writel (readl(base + EMAC_RCNTRL_REG) | (JUMBO_FRAME_SIZE << 16), base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC enable stacked vlan function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_stacked_vlan(void *base) ++{ ++ /* MTIP doesn't support stacked vlan */ ++ return; ++} ++ ++/** GEMAC enable pause rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_enable_pause_rx(void *base) ++{ ++ writel (readl(base + EMAC_RCNTRL_REG) | EMAC_RCNTRL_FCE, base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC disable pause rx function. ++* @param[in] base GEMAC base address ++*/ ++void gemac_disable_pause_rx(void *base) ++{ ++ writel (readl(base + EMAC_RCNTRL_REG) & ~EMAC_RCNTRL_FCE, base + EMAC_RCNTRL_REG); ++} ++ ++/** GEMAC wol configuration ++* @param[in] base GEMAC base address ++* @param[in] wol_conf WoL register configuration ++*/ ++void gemac_set_wol(void *base, u32 wol_conf) ++{ ++ printk("%s() TODO\n", __func__); ++} ++ ++/** Sets Gemac bus width to 64bit ++ * @param[in] base GEMAC base address ++ * @param[in] width gemac bus width to be set possible values are 32/64/128 ++ * */ ++void gemac_set_bus_width(void *base, int width) ++{ ++} ++ ++/** Sets Gemac configuration. ++* @param[in] base GEMAC base address ++* @param[in] cfg GEMAC configuration ++*/ ++void gemac_set_config(void *base, GEMAC_CFG *cfg) ++{ ++ ++ /*GEMAC config taken from VLSI */ ++ writel(0x00000004, base + EMAC_TFWR_STR_FWD); ++ writel(0x00000005, base + EMAC_RX_SECTIOM_FULL); ++ writel(0x00003fff, base + EMAC_TRUNC_FL); ++ writel(0x00000030, base + EMAC_TX_SECTION_EMPTY); ++ writel(0x00000000, base + EMAC_MIB_CTRL_STS_REG); ++ ++ gemac_set_mode(base, cfg->mode); ++ ++ gemac_set_speed(base, cfg->speed); ++ ++ gemac_set_duplex(base,cfg->duplex); ++} ++ ++ ++#endif //CONFIG_PLATFORM_LS1012A) ++ ++ ++ ++/**************************** GPI ***************************/ ++ ++/** Initializes a GPI block. ++* @param[in] base GPI base address ++* @param[in] cfg GPI configuration ++*/ ++void gpi_init(void *base, GPI_CFG *cfg) ++{ ++ gpi_reset(base); ++ ++ gpi_disable(base); ++ ++ gpi_set_config(base, cfg); ++} ++ ++/** Resets a GPI block. ++* @param[in] base GPI base address ++*/ ++void gpi_reset(void *base) ++{ ++ writel (CORE_SW_RESET, base + GPI_CTRL); ++} ++ ++/** Enables a GPI block. ++* @param[in] base GPI base address ++*/ ++void gpi_enable(void *base) ++{ ++ writel (CORE_ENABLE, base + GPI_CTRL); ++} ++ ++/** Disables a GPI block. ++* @param[in] base GPI base address ++*/ ++void gpi_disable(void *base) ++{ ++ writel (CORE_DISABLE, base + GPI_CTRL); ++} ++ ++ ++/** Sets the configuration of a GPI block. ++* @param[in] base GPI base address ++* @param[in] cfg GPI configuration ++*/ ++void gpi_set_config(void *base, GPI_CFG *cfg) ++{ ++ writel (CBUS_VIRT_TO_PFE(BMU1_BASE_ADDR + BMU_ALLOC_CTRL), base + GPI_LMEM_ALLOC_ADDR); ++ writel (CBUS_VIRT_TO_PFE(BMU1_BASE_ADDR + BMU_FREE_CTRL), base + GPI_LMEM_FREE_ADDR); ++ writel (CBUS_VIRT_TO_PFE(BMU2_BASE_ADDR + BMU_ALLOC_CTRL), base + GPI_DDR_ALLOC_ADDR); ++ writel (CBUS_VIRT_TO_PFE(BMU2_BASE_ADDR + BMU_FREE_CTRL), base + GPI_DDR_FREE_ADDR); ++ writel (CBUS_VIRT_TO_PFE(CLASS_INQ_PKTPTR), base + GPI_CLASS_ADDR); ++ writel (DDR_HDR_SIZE, base + GPI_DDR_DATA_OFFSET); ++ writel (LMEM_HDR_SIZE, base + GPI_LMEM_DATA_OFFSET); ++ writel (0, base + GPI_LMEM_SEC_BUF_DATA_OFFSET); ++ writel (0, base + GPI_DDR_SEC_BUF_DATA_OFFSET); ++ writel ((DDR_HDR_SIZE << 16) | LMEM_HDR_SIZE, base + GPI_HDR_SIZE); ++ writel ((DDR_BUF_SIZE << 16) | LMEM_BUF_SIZE, base + GPI_BUF_SIZE); ++ ++ writel (((cfg->lmem_rtry_cnt << 16) | (GPI_DDR_BUF_EN << 1) | GPI_LMEM_BUF_EN), base + GPI_RX_CONFIG); ++ writel (cfg->tmlf_txthres, base + GPI_TMLF_TX); ++ writel (cfg->aseq_len, base + GPI_DTX_ASEQ); ++ writel (1, base + GPI_TOE_CHKSUM_EN); ++} ++ ++/**************************** CLASSIFIER ***************************/ ++ ++/** Initializes CLASSIFIER block. ++* @param[in] cfg CLASSIFIER configuration ++*/ ++void class_init(CLASS_CFG *cfg) ++{ ++ class_reset(); ++ ++ class_disable(); ++ ++ class_set_config(cfg); ++} ++ ++/** Resets CLASSIFIER block. ++* ++*/ ++void class_reset(void) ++{ ++ writel(CORE_SW_RESET, CLASS_TX_CTRL); ++} ++ ++/** Enables all CLASS-PE's cores. ++* ++*/ ++void class_enable(void) ++{ ++ writel(CORE_ENABLE, CLASS_TX_CTRL); ++} ++ ++/** Disables all CLASS-PE's cores. ++* ++*/ ++void class_disable(void) ++{ ++ writel(CORE_DISABLE, CLASS_TX_CTRL); ++} ++ ++/** Sets the configuration of the CLASSIFIER block. ++* @param[in] cfg CLASSIFIER configuration ++*/ ++void class_set_config(CLASS_CFG *cfg) ++{ ++ u32 val; ++ ++ /* Initialize route table */ ++ if (!cfg->resume) ++ memset(DDR_PHYS_TO_VIRT(cfg->route_table_baseaddr), 0, (1 << cfg->route_table_hash_bits) * CLASS_ROUTE_SIZE); ++ ++#if !defined(LS1012A_PFE_RESET_WA) ++ writel(cfg->pe_sys_clk_ratio, CLASS_PE_SYS_CLK_RATIO); ++#endif ++ ++ writel((DDR_HDR_SIZE << 16) | LMEM_HDR_SIZE, CLASS_HDR_SIZE); ++ writel(LMEM_BUF_SIZE, CLASS_LMEM_BUF_SIZE); ++ writel(CLASS_ROUTE_ENTRY_SIZE(CLASS_ROUTE_SIZE) | CLASS_ROUTE_HASH_SIZE(cfg->route_table_hash_bits), CLASS_ROUTE_HASH_ENTRY_SIZE); ++ writel(HIF_PKT_CLASS_EN| HIF_PKT_OFFSET(sizeof(struct hif_hdr)), CLASS_HIF_PARSE); ++ ++ val = HASH_CRC_PORT_IP | QB2BUS_LE; ++ ++#if defined(CONFIG_IP_ALIGNED) ++ val |= IP_ALIGNED; ++#endif ++ ++ /* Class PE packet steering will only work if TOE mode, bridge fetch or ++ * route fetch are enabled (see class/qb_fet.v). Route fetch would trigger ++ * additional memory copies (likely from DDR because of hash table size, which ++ * cannot be reduced because PE software still relies on hash value computed ++ * in HW), so when not in TOE mode we simply enable HW bridge fetch even ++ * though we don't use it. ++ */ ++ if (cfg->toe_mode) ++ val |= CLASS_TOE; ++ else ++ val |= HW_BRIDGE_FETCH; ++ ++ writel(val, CLASS_ROUTE_MULTI); ++ ++ writel(DDR_PHYS_TO_PFE(cfg->route_table_baseaddr), CLASS_ROUTE_TABLE_BASE); ++ writel(CLASS_PE0_RO_DM_ADDR0_VAL, CLASS_PE0_RO_DM_ADDR0); ++ writel(CLASS_PE0_RO_DM_ADDR1_VAL, CLASS_PE0_RO_DM_ADDR1); ++ writel(CLASS_PE0_QB_DM_ADDR0_VAL, CLASS_PE0_QB_DM_ADDR0); ++ writel(CLASS_PE0_QB_DM_ADDR1_VAL, CLASS_PE0_QB_DM_ADDR1); ++ writel(CBUS_VIRT_TO_PFE(TMU_PHY_INQ_PKTPTR), CLASS_TM_INQ_ADDR); ++ ++ writel(23, CLASS_AFULL_THRES); ++ writel(23, CLASS_TSQ_FIFO_THRES); ++ ++ writel(24, CLASS_MAX_BUF_CNT); ++ writel(24, CLASS_TSQ_MAX_CNT); ++} ++ ++/**************************** TMU ***************************/ ++ ++void tmu_reset(void) ++{ ++ writel(SW_RESET, TMU_CTRL); ++} ++ ++/** Initializes TMU block. ++* @param[in] cfg TMU configuration ++*/ ++void tmu_init(TMU_CFG *cfg) ++{ ++ int q, phyno; ++ ++ tmu_disable(0xF); ++ mdelay(10); ++ ++#if !defined(LS1012A_PFE_RESET_WA) ++ /* keep in soft reset */ ++ writel(SW_RESET, TMU_CTRL); ++#endif ++ writel(0x3, TMU_SYS_GENERIC_CONTROL); ++ writel(750, TMU_INQ_WATERMARK); ++ writel(CBUS_VIRT_TO_PFE(EGPI1_BASE_ADDR + GPI_INQ_PKTPTR), TMU_PHY0_INQ_ADDR); ++ writel(CBUS_VIRT_TO_PFE(EGPI2_BASE_ADDR + GPI_INQ_PKTPTR), TMU_PHY1_INQ_ADDR); ++#if !defined(CONFIG_PLATFORM_LS1012A) ++ writel(CBUS_VIRT_TO_PFE(EGPI3_BASE_ADDR + GPI_INQ_PKTPTR), TMU_PHY2_INQ_ADDR); ++#endif ++ writel(CBUS_VIRT_TO_PFE(HGPI_BASE_ADDR + GPI_INQ_PKTPTR), TMU_PHY3_INQ_ADDR); ++ writel(CBUS_VIRT_TO_PFE(HIF_NOCPY_RX_INQ0_PKTPTR), TMU_PHY4_INQ_ADDR); ++ writel(CBUS_VIRT_TO_PFE(UTIL_INQ_PKTPTR), TMU_PHY5_INQ_ADDR); ++ writel(CBUS_VIRT_TO_PFE(BMU2_BASE_ADDR + BMU_FREE_CTRL), TMU_BMU_INQ_ADDR); ++ ++ writel(0x3FF, TMU_TDQ0_SCH_CTRL); // enabling all 10 schedulers [9:0] of each TDQ ++ writel(0x3FF, TMU_TDQ1_SCH_CTRL); ++#if !defined(CONFIG_PLATFORM_LS1012A) ++ writel(0x3FF, TMU_TDQ2_SCH_CTRL); ++#endif ++ writel(0x3FF, TMU_TDQ3_SCH_CTRL); ++ ++#if !defined(LS1012A_PFE_RESET_WA) ++ writel(cfg->pe_sys_clk_ratio, TMU_PE_SYS_CLK_RATIO); ++#endif ++ ++#if !defined(LS1012A_PFE_RESET_WA) ++ writel(DDR_PHYS_TO_PFE(cfg->llm_base_addr), TMU_LLM_BASE_ADDR); // Extra packet pointers will be stored from this address onwards ++ ++ writel(cfg->llm_queue_len, TMU_LLM_QUE_LEN); ++ writel(5, TMU_TDQ_IIFG_CFG); ++ writel(DDR_BUF_SIZE, TMU_BMU_BUF_SIZE); ++ ++ writel(0x0, TMU_CTRL); ++ ++ /* MEM init */ ++ printk(KERN_INFO "%s: mem init\n", __func__); ++ writel(MEM_INIT, TMU_CTRL); ++ ++ while(!(readl(TMU_CTRL) & MEM_INIT_DONE)) ; ++ ++ /* LLM init */ ++ printk(KERN_INFO "%s: lmem init\n", __func__); ++ writel(LLM_INIT, TMU_CTRL); ++ ++ while(!(readl(TMU_CTRL) & LLM_INIT_DONE)) ; ++#endif ++ // set up each queue for tail drop ++ for (phyno = 0; phyno < 4; phyno++) ++ { ++#if defined(CONFIG_PLATFORM_LS1012A) ++ if(phyno == 2) continue; ++#endif ++ for (q = 0; q < 16; q++) ++ { ++ u32 qdepth; ++ writel((phyno << 8) | q, TMU_TEQ_CTRL); ++ writel(1 << 22, TMU_TEQ_QCFG); //Enable tail drop ++ ++ if (phyno == 3) ++ qdepth = DEFAULT_TMU3_QDEPTH; ++ else ++ qdepth = (q == 0) ? DEFAULT_Q0_QDEPTH : DEFAULT_MAX_QDEPTH; ++ ++ // LOG: 68855 ++ // The following is a workaround for the reordered packet and BMU2 buffer leakage issue. ++ if (CHIP_REVISION() == 0) ++ qdepth = 31; ++ ++ writel(qdepth << 18, TMU_TEQ_HW_PROB_CFG2); ++ writel(qdepth >> 14, TMU_TEQ_HW_PROB_CFG3); ++ } ++ } ++ ++#ifdef CFG_LRO ++ /* Set TMU-3 queue 5 (LRO) in no-drop mode */ ++ writel((3 << 8) | TMU_QUEUE_LRO, TMU_TEQ_CTRL); ++ writel(0, TMU_TEQ_QCFG); ++#endif ++ ++ writel(0x05, TMU_TEQ_DISABLE_DROPCHK); ++ ++ writel(0x0, TMU_CTRL); ++} ++ ++/** Enables TMU-PE cores. ++* @param[in] pe_mask TMU PE mask ++*/ ++void tmu_enable(u32 pe_mask) ++{ ++ writel(readl(TMU_TX_CTRL) | (pe_mask & 0xF), TMU_TX_CTRL); ++} ++ ++/** Disables TMU cores. ++* @param[in] pe_mask TMU PE mask ++*/ ++void tmu_disable(u32 pe_mask) ++{ ++ writel(readl(TMU_TX_CTRL) & ~(pe_mask & 0xF), TMU_TX_CTRL); ++} ++/** This will return the tmu queue status ++ * @param[in] if_id gem interface id or TMU index ++ * @return returns the bit mask of busy queues, zero means all queues are empty ++ */ ++u32 tmu_qstatus(u32 if_id) ++{ ++ return cpu_to_be32(pe_dmem_read(TMU0_ID+if_id, PESTATUS_ADDR_TMU + offsetof(PE_STATUS, tmu_qstatus), 4)); ++} ++ ++u32 tmu_pkts_processed(u32 if_id) ++{ ++ return cpu_to_be32(pe_dmem_read(TMU0_ID+if_id, PESTATUS_ADDR_TMU + offsetof(PE_STATUS, rx), 4)); ++} ++/**************************** UTIL ***************************/ ++ ++/** Resets UTIL block. ++*/ ++void util_reset(void) ++{ ++ writel(CORE_SW_RESET, UTIL_TX_CTRL); ++} ++ ++/** Initializes UTIL block. ++* @param[in] cfg UTIL configuration ++*/ ++void util_init(UTIL_CFG *cfg) ++{ ++ writel(cfg->pe_sys_clk_ratio, UTIL_PE_SYS_CLK_RATIO); ++} ++ ++/** Enables UTIL-PE core. ++* ++*/ ++void util_enable(void) ++{ ++ writel(CORE_ENABLE, UTIL_TX_CTRL); ++} ++ ++/** Disables UTIL-PE core. ++* ++*/ ++void util_disable(void) ++{ ++ writel(CORE_DISABLE, UTIL_TX_CTRL); ++} ++ ++/**************************** HIF ***************************/ ++ ++/** Initializes HIF no copy block. ++* ++*/ ++void hif_nocpy_init(void) ++{ ++ writel(4, HIF_NOCPY_TX_PORT_NO); ++ writel(CBUS_VIRT_TO_PFE(BMU1_BASE_ADDR + BMU_ALLOC_CTRL), HIF_NOCPY_LMEM_ALLOC_ADDR); ++ writel(CBUS_VIRT_TO_PFE(CLASS_INQ_PKTPTR), HIF_NOCPY_CLASS_ADDR); ++ writel(CBUS_VIRT_TO_PFE(TMU_PHY_INQ_PKTPTR), HIF_NOCPY_TMU_PORT0_ADDR); ++ writel(HIF_RX_POLL_CTRL_CYCLE<<16|HIF_TX_POLL_CTRL_CYCLE, HIF_NOCPY_POLL_CTRL); ++} ++ ++/** Enable hif_nocpy tx DMA and interrupt ++* ++*/ ++void hif_nocpy_tx_enable(void) ++{ ++ /*TODO not sure poll_cntrl_en is required or not */ ++ writel( HIF_CTRL_DMA_EN, HIF_NOCPY_TX_CTRL); ++ //writel((readl(HIF_NOCPY_INT_ENABLE) | HIF_INT_EN | HIF_TXPKT_INT_EN), HIF_NOCPY_INT_ENABLE); ++} ++ ++/** Disable hif_nocpy tx DMA and interrupt ++* ++*/ ++void hif_nocpy_tx_disable(void) ++{ ++ u32 hif_int; ++ ++ writel(0, HIF_NOCPY_TX_CTRL); ++ ++ hif_int = readl(HIF_NOCPY_INT_ENABLE); ++ hif_int &= HIF_TXPKT_INT_EN; ++ writel(hif_int, HIF_NOCPY_INT_ENABLE); ++} ++ ++/** Enable hif rx DMA and interrupt ++* ++*/ ++void hif_nocpy_rx_enable(void) ++{ ++ hif_nocpy_rx_dma_start(); ++ writel((readl(HIF_NOCPY_INT_ENABLE) | HIF_INT_EN | HIF_RXPKT_INT_EN), HIF_NOCPY_INT_ENABLE); ++} ++ ++/** Disable hif_nocpy rx DMA and interrupt ++* ++*/ ++void hif_nocpy_rx_disable(void) ++{ ++ u32 hif_int; ++ ++ writel(0, HIF_NOCPY_RX_CTRL); ++ ++ hif_int = readl(HIF_NOCPY_INT_ENABLE); ++ hif_int &= HIF_RXPKT_INT_EN; ++ writel(hif_int, HIF_NOCPY_INT_ENABLE); ++ ++} ++/** Initializes HIF copy block. ++* ++*/ ++void hif_init(void) ++{ ++ /*Initialize HIF registers*/ ++ writel((HIF_RX_POLL_CTRL_CYCLE << 16) | HIF_TX_POLL_CTRL_CYCLE, HIF_POLL_CTRL); ++} ++ ++/** Enable hif tx DMA and interrupt ++* ++*/ ++void hif_tx_enable(void) ++{ ++ /*TODO not sure poll_cntrl_en is required or not */ ++ writel(HIF_CTRL_DMA_EN, HIF_TX_CTRL); ++ writel((readl(HIF_INT_ENABLE) | HIF_INT_EN | HIF_TXPKT_INT_EN), HIF_INT_ENABLE); ++} ++ ++/** Disable hif tx DMA and interrupt ++* ++*/ ++void hif_tx_disable(void) ++{ ++ u32 hif_int; ++ ++ writel(0, HIF_TX_CTRL); ++ ++ hif_int = readl(HIF_INT_ENABLE); ++ hif_int &= HIF_TXPKT_INT_EN; ++ writel(hif_int, HIF_INT_ENABLE); ++} ++ ++/** Enable hif rx DMA and interrupt ++* ++*/ ++void hif_rx_enable(void) ++{ ++ hif_rx_dma_start(); ++ writel((readl(HIF_INT_ENABLE) | HIF_INT_EN | HIF_RXPKT_INT_EN), HIF_INT_ENABLE); ++} ++ ++/** Disable hif rx DMA and interrupt ++* ++*/ ++void hif_rx_disable(void) ++{ ++ u32 hif_int; ++ ++ writel(0, HIF_RX_CTRL); ++ ++ hif_int = readl(HIF_INT_ENABLE); ++ hif_int &= HIF_RXPKT_INT_EN; ++ writel(hif_int, HIF_INT_ENABLE); ++ ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hif.c +@@ -0,0 +1,939 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifdef __KERNEL__ ++#include <linux/kernel.h> ++#include <linux/interrupt.h> ++#include <linux/dma-mapping.h> ++#include <linux/dmapool.h> ++#include <linux/sched.h> ++#include <linux/module.h> ++#include <linux/list.h> ++#include <linux/kthread.h> ++#include <linux/slab.h> ++ ++#include <asm/io.h> ++#include <asm/irq.h> ++#else ++#include "platform.h" ++#endif ++ ++ ++#include "pfe_mod.h" ++#if 0 ++#define DMA_MAP_SINGLE(dev, vaddr, size, direction) dma_map_single(dev, vaddr, size, direction) ++#define DMA_UNMAP_SINGLE(dev, vaddr, size, direction) dma_unmap_single(dev, vaddr, size, direction) ++void ct_flush(void *addr, u32 size) ++{ ++ dma_map_single(pfe->dev, addr, size, DMA_TO_DEVICE); ++} ++#else ++#define DMA_UNMAP_SINGLE(dev, vaddr, size, direction) ++#define DMA_MAP_SINGLE(dev, vaddr, size, direction) virt_to_phys(vaddr) ++#define ct_flush(addr, sz) ++#endif ++ ++#define HIF_INT_MASK (HIF_INT | HIF_RXPKT_INT) ++ ++#define inc_cl_idx(idxname) idxname = (idxname+1) & (queue->size-1) ++#define inc_hif_rxidx(idxname) idxname = (idxname+1) & (hif->RxRingSize-1) ++#define inc_hif_txidx(idxname) idxname = (idxname+1) & (hif->TxRingSize-1) ++ ++unsigned char napi_first_batch = 0; ++ ++static int pfe_hif_alloc_descr(struct pfe_hif *hif) ++{ ++#if !defined(CONFIG_PLATFORM_PCI) ++ void *addr; ++ dma_addr_t dma_addr; ++ int err = 0; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ addr = dma_alloc_coherent(pfe->dev, ++ HIF_RX_DESC_NT * sizeof(struct hif_desc) + HIF_TX_DESC_NT * sizeof(struct hif_desc), ++ &dma_addr, GFP_KERNEL); ++ ++ if (!addr) { ++ printk(KERN_ERR "%s: Could not allocate buffer descriptors!\n", __func__); ++ err = -ENOMEM; ++ goto err0; ++ } ++ ++ hif->descr_baseaddr_p = dma_addr; ++ hif->descr_baseaddr_v = addr; ++#else ++ hif->descr_baseaddr_p = pfe->ddr_phys_baseaddr + HIF_DESC_BASEADDR; ++ hif->descr_baseaddr_v = pfe->ddr_baseaddr + HIF_DESC_BASEADDR; ++#endif ++ hif->RxRingSize = HIF_RX_DESC_NT; ++ hif->TxRingSize = HIF_TX_DESC_NT; ++ ++ return 0; ++ ++err0: ++ return err; ++} ++ ++static void pfe_hif_free_descr(struct pfe_hif *hif) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++#if !defined(CONFIG_PLATFORM_PCI) ++ dma_free_coherent(pfe->dev, ++ hif->RxRingSize * sizeof(struct hif_desc) + hif->TxRingSize * sizeof(struct hif_desc), ++ hif->descr_baseaddr_v, hif->descr_baseaddr_p); ++#endif ++} ++void pfe_hif_desc_dump(struct pfe_hif *hif) ++{ ++ struct hif_desc *desc; ++ unsigned long desc_p; ++ int ii=0; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ desc = hif->RxBase; ++ desc_p = (u32)((u64)desc - (u64)hif->descr_baseaddr_v + hif->descr_baseaddr_p); ++ ++ printk("HIF Rx desc base %p physical %x\n", desc, (u32)desc_p); ++ for (ii = 0; ii < hif->RxRingSize; ii++) { ++ printk(KERN_INFO "status: %08x, ctrl: %08x, data: %08x, next: %x\n", ++ desc->status, desc->ctrl, desc->data, desc->next); ++ desc++; ++ } ++ ++ desc = hif->TxBase; ++ desc_p = ((u64)desc - (u64)hif->descr_baseaddr_v + hif->descr_baseaddr_p); ++ ++ printk("HIF Tx desc base %p physical %x\n", desc, (u32)desc_p); ++ for (ii = 0; ii < hif->TxRingSize; ii++) { ++ printk(KERN_INFO "status: %08x, ctrl: %08x, data: %08x, next: %x\n", ++ desc->status, desc->ctrl, desc->data, desc->next); ++ desc++; ++ } ++ ++} ++ ++/* pfe_hif_release_buffers ++ * ++ */ ++static void pfe_hif_release_buffers(struct pfe_hif *hif) ++{ ++ struct hif_desc *desc; ++ int i = 0; ++ ++ hif->RxBase = hif->descr_baseaddr_v; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ /*Free Rx buffers */ ++#if !defined(CONFIG_PLATFORM_PCI) ++ desc = hif->RxBase; ++ for (i = 0; i < hif->RxRingSize; i++) { ++ if (desc->data) { ++ if ((i < hif->shm->rx_buf_pool_cnt) && (hif->shm->rx_buf_pool[i] == NULL)) { ++ //dma_unmap_single(hif->dev, desc->data, hif->rx_buf_len[i], DMA_FROM_DEVICE); ++ DMA_UNMAP_SINGLE(hif->dev, desc->data, hif->rx_buf_len[i], DMA_FROM_DEVICE); ++ hif->shm->rx_buf_pool[i] = hif->rx_buf_addr[i]; ++ } ++ else { ++ /*TODO This should not happen*/ ++ printk(KERN_ERR "%s: buffer pool already full\n", __func__); ++ } ++ } ++ ++ desc->data = 0; ++ desc->status = 0; ++ desc->ctrl = 0; ++ desc++; ++ } ++#endif ++} ++ ++ ++/* ++ * pfe_hif_init_buffers ++ * This function initializes the HIF Rx/Tx ring descriptors and ++ * initialize Rx queue with buffers. ++ */ ++static int pfe_hif_init_buffers(struct pfe_hif *hif) ++{ ++ struct hif_desc *desc, *first_desc_p; ++ u32 data; ++ int i = 0; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ /* Check enough Rx buffers available in the shared memory */ ++ if (hif->shm->rx_buf_pool_cnt < hif->RxRingSize) ++ return -ENOMEM; ++ ++ hif->RxBase = hif->descr_baseaddr_v; ++ memset(hif->RxBase, 0, hif->RxRingSize * sizeof(struct hif_desc)); ++ ++ /*Initialize Rx descriptors */ ++ desc = hif->RxBase; ++ first_desc_p = (struct hif_desc *)hif->descr_baseaddr_p; ++ ++ for (i = 0; i < hif->RxRingSize; i++) { ++ /* Initialize Rx buffers from the shared memory */ ++ ++#if defined(CONFIG_PLATFORM_PCI) ++ data = pfe->ddr_phys_baseaddr + HIF_RX_PKT_DDR_BASEADDR + i * DDR_BUF_SIZE; ++#else ++ data = (u32)DMA_MAP_SINGLE(hif->dev, hif->shm->rx_buf_pool[i], pfe_pkt_size, DMA_FROM_DEVICE); ++ hif->rx_buf_addr[i] = hif->shm->rx_buf_pool[i]; ++ hif->rx_buf_len[i] = pfe_pkt_size; ++ // printk("#%d %p %p %d\n", i, data, hif->rx_buf_addr[i], hif->rx_buf_len[i]); ++ hif->shm->rx_buf_pool[i] = NULL; ++#endif ++ if (likely(dma_mapping_error(hif->dev, data) == 0)) { ++ desc->data = DDR_PHYS_TO_PFE(data); ++ } else { ++ printk(KERN_ERR "%s : low on mem\n", __func__); ++ ++ goto err; ++ } ++ ++ desc->status = 0; ++ wmb(); ++ desc->ctrl = BD_CTRL_PKT_INT_EN | BD_CTRL_LIFM | BD_CTRL_DIR | ++ BD_CTRL_DESC_EN | BD_BUF_LEN(pfe_pkt_size); ++ /* Chain descriptors */ ++ desc->next = (u32)DDR_PHYS_TO_PFE(first_desc_p + i + 1); ++ desc++; ++ } ++ ++ /* Overwrite last descriptor to chain it to first one*/ ++ desc--; ++ desc->next = (u32)DDR_PHYS_TO_PFE(first_desc_p); ++ ++ hif->RxtocleanIndex = 0; ++ ++ /*Initialize Rx buffer descriptor ring base address */ ++ writel(DDR_PHYS_TO_PFE(hif->descr_baseaddr_p), HIF_RX_BDP_ADDR); ++ ++ hif->TxBase = hif->RxBase + hif->RxRingSize; ++ first_desc_p = (struct hif_desc *)hif->descr_baseaddr_p + hif->RxRingSize; ++ memset(hif->TxBase, 0, hif->TxRingSize * sizeof(struct hif_desc)); ++ ++ /*Initialize tx descriptors */ ++ desc = hif->TxBase; ++ ++ for (i = 0; i < hif->TxRingSize; i++) { ++ /* Chain descriptors */ ++ desc->next = (u32)DDR_PHYS_TO_PFE(first_desc_p + i + 1); ++#if defined(CONFIG_PLATFORM_PCI) ++ desc->data = pfe->ddr_phys_baseaddr + HIF_TX_PKT_DDR_BASEADDR + i * DDR_BUF_SIZE; ++#endif ++ desc->ctrl = 0; ++ desc++; ++ } ++ ++ /* Overwrite last descriptor to chain it to first one */ ++ desc--; ++ desc->next = (u32)DDR_PHYS_TO_PFE(first_desc_p); ++ hif->TxAvail = hif->TxRingSize; ++ hif->Txtosend = 0; ++ hif->Txtoclean = 0; ++ hif->Txtoflush = 0; ++ ++ /*Initialize Tx buffer descriptor ring base address */ ++ writel((u32)DDR_PHYS_TO_PFE(first_desc_p), HIF_TX_BDP_ADDR); ++ ++ return 0; ++ ++err: ++ pfe_hif_release_buffers(hif); ++ return -ENOMEM; ++} ++ ++/* pfe_hif_client_register ++ * ++ * This function used to register a client driver with the HIF driver. ++ * ++ * Return value: ++ * 0 - on Successful registration ++ */ ++static int pfe_hif_client_register(struct pfe_hif *hif, u32 client_id, struct hif_client_shm *client_shm) ++{ ++ struct hif_client *client = &hif->client[client_id]; ++ u32 i, cnt; ++ struct rx_queue_desc *rx_qbase; ++ struct tx_queue_desc *tx_qbase; ++ struct hif_rx_queue *rx_queue; ++ struct hif_tx_queue *tx_queue; ++ int err = 0; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ spin_lock_bh(&hif->tx_lock); ++ ++ if (test_bit(client_id, &hif->shm->gClient_status[0])) { ++ printk(KERN_ERR "%s: client %d already registered\n", __func__, client_id); ++ err = -1; ++ goto unlock; ++ } ++ ++ memset(client, 0, sizeof(struct hif_client)); ++ ++ /*Initialize client Rx queues baseaddr, size */ ++ ++ cnt = CLIENT_CTRL_RX_Q_CNT(client_shm->ctrl); ++ /*Check if client is requesting for more queues than supported */ ++ if (cnt > HIF_CLIENT_QUEUES_MAX) ++ cnt = HIF_CLIENT_QUEUES_MAX; ++ ++ client->rx_qn = cnt; ++ rx_qbase = (struct rx_queue_desc *)client_shm->rx_qbase; ++ for (i = 0; i < cnt; i++) ++ { ++ rx_queue = &client->rx_q[i]; ++ rx_queue->base = rx_qbase + i * client_shm->rx_qsize; ++ rx_queue->size = client_shm->rx_qsize; ++ rx_queue->write_idx = 0; ++ } ++ ++ /*Initialize client Tx queues baseaddr, size */ ++ cnt = CLIENT_CTRL_TX_Q_CNT(client_shm->ctrl); ++ ++ /*Check if client is requesting for more queues than supported */ ++ if (cnt > HIF_CLIENT_QUEUES_MAX) ++ cnt = HIF_CLIENT_QUEUES_MAX; ++ ++ client->tx_qn = cnt; ++ tx_qbase = (struct tx_queue_desc *)client_shm->tx_qbase; ++ for (i = 0; i < cnt; i++) ++ { ++ tx_queue = &client->tx_q[i]; ++ tx_queue->base = tx_qbase + i * client_shm->tx_qsize; ++ tx_queue->size = client_shm->tx_qsize; ++ tx_queue->ack_idx = 0; ++ } ++ ++ set_bit(client_id, &hif->shm->gClient_status[0]); ++ ++unlock: ++ spin_unlock_bh(&hif->tx_lock); ++ ++ return err; ++} ++ ++ ++/* pfe_hif_client_unregister ++ * ++ * This function used to unregister a client from the HIF driver. ++ * ++ */ ++static void pfe_hif_client_unregister(struct pfe_hif *hif, u32 client_id) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ /* Mark client as no longer available (which prevents further packet receive for this client) */ ++ spin_lock_bh(&hif->tx_lock); ++ ++ if (!test_bit(client_id, &hif->shm->gClient_status[0])) { ++ printk(KERN_ERR "%s: client %d not registered\n", __func__, client_id); ++ ++ spin_unlock_bh(&hif->tx_lock); ++ return; ++ } ++ ++ clear_bit(client_id, &hif->shm->gClient_status[0]); ++ ++ spin_unlock_bh(&hif->tx_lock); ++} ++ ++/* client_put_rxpacket- ++ * This functions puts the Rx pkt in the given client Rx queue. ++ * It actually swap the Rx pkt in the client Rx descriptor buffer ++ * and returns the free buffer from it. ++ * ++ * If the funtion returns NULL means client Rx queue is full and ++ * packet couldn't send to client queue. ++ */ ++static void *client_put_rxpacket(struct hif_rx_queue *queue, void *pkt, u32 len, u32 flags, u32 client_ctrl, u32 *rem_len) ++{ ++ void *free_pkt = NULL; ++ struct rx_queue_desc *desc = queue->base + queue->write_idx; ++ ++ if (desc->ctrl & CL_DESC_OWN) { ++#if defined(CONFIG_PLATFORM_PCI) ++ memcpy(desc->data, pkt, len); ++ free_pkt = PFE_HOST_TO_PCI(pkt); ++ smp_wmb(); ++ desc->ctrl = CL_DESC_BUF_LEN(len) | flags; ++ inc_cl_idx(queue->write_idx); ++#else ++ //TODO: move allocations after Rx loop to improve instruction cache locality ++ if (page_mode) { ++ int rem_page_size = PAGE_SIZE - PRESENT_OFST_IN_PAGE(pkt); ++ int cur_pkt_size = ROUND_MIN_RX_SIZE(len + pfe_pkt_headroom); ++ *rem_len = (rem_page_size - cur_pkt_size); ++ //printk("%p rem_len %d cur_len %d buf_len %d\n", pkt, rem_page_size, cur_pkt_size, *rem_len); ++ if (*rem_len) ++ { ++ free_pkt = pkt + cur_pkt_size; ++ get_page(virt_to_page(free_pkt)); ++ } else { ++ free_pkt = (void *)__get_free_page(GFP_ATOMIC | GFP_DMA_PFE); ++ *rem_len = pfe_pkt_size; ++ } ++ } else { ++ free_pkt = kmalloc(PFE_BUF_SIZE, GFP_ATOMIC | GFP_DMA_PFE); ++ *rem_len = PFE_BUF_SIZE - pfe_pkt_headroom; ++ } ++ ++ if (free_pkt) { ++ desc->data = pkt; ++ desc->client_ctrl = client_ctrl; ++ smp_wmb(); ++ desc->ctrl = CL_DESC_BUF_LEN(len) | flags; ++ inc_cl_idx(queue->write_idx); ++ free_pkt += pfe_pkt_headroom; ++ } ++#endif ++ } ++ ++ return free_pkt; ++} ++ ++ ++/* pfe_hif_rx_process- ++ * This function does pfe hif rx queue processing. ++ * Dequeue packet from Rx queue and send it to corresponding client queue ++ */ ++static int pfe_hif_rx_process(struct pfe_hif *hif, int budget) ++{ ++ struct hif_desc *desc; ++ struct hif_hdr *pkt_hdr; ++ struct __hif_hdr hif_hdr; ++ void *free_buf; ++ int rtc, len, rx_processed = 0; ++ struct __hif_desc local_desc; ++ int flags; ++ unsigned int desc_p; ++ unsigned int buf_size = 0; ++ ++ spin_lock_bh(&hif->lock); ++ ++ rtc = hif->RxtocleanIndex; ++ ++ while (rx_processed < budget) ++ { ++ /*TODO may need to implement rx process budget */ ++ desc = hif->RxBase + rtc; ++ ++ __memcpy12(&local_desc, desc); ++ ++ /* ACK pending Rx interrupt */ ++ if (local_desc.ctrl & BD_CTRL_DESC_EN) { ++ writel(HIF_INT_MASK, HIF_INT_SRC); ++ ++ if(rx_processed == 0) ++ { ++ if(napi_first_batch == 1) ++ { ++ desc_p = hif->descr_baseaddr_p + ((unsigned long int)(desc) - (unsigned long int)hif->descr_baseaddr_v); ++#if defined(CONFIG_PLATFORM_C2000) ++ outer_inv_range(desc_p, (desc_p + 16)); ++#endif ++ napi_first_batch = 0; ++ } ++ } ++ ++ __memcpy12(&local_desc, desc); ++ ++ if (local_desc.ctrl & BD_CTRL_DESC_EN) ++ break; ++ } ++ ++ napi_first_batch = 0; ++ ++#ifdef HIF_NAPI_STATS ++ hif->napi_counters[NAPI_DESC_COUNT]++; ++#endif ++ len = BD_BUF_LEN(local_desc.ctrl); ++#if defined(CONFIG_PLATFORM_PCI) ++ pkt_hdr = &hif_hdr; ++ memcpy(pkt_hdr, (void *)PFE_PCI_TO_HOST(local_desc.data), sizeof(struct hif_hdr)); ++#else ++ //dma_unmap_single(hif->dev, DDR_PFE_TO_PHYS(local_desc.data), hif->rx_buf_len[rtc], DMA_FROM_DEVICE); ++ DMA_UNMAP_SINGLE(hif->dev, DDR_PFE_TO_PHYS(local_desc.data), hif->rx_buf_len[rtc], DMA_FROM_DEVICE); ++ ++ pkt_hdr = (struct hif_hdr *)hif->rx_buf_addr[rtc]; ++ ++ /* Track last HIF header received */ ++ if (!hif->started) { ++ hif->started = 1; ++ ++ __memcpy8(&hif_hdr, pkt_hdr); ++ ++ hif->qno = hif_hdr.hdr.qNo; ++ hif->client_id = hif_hdr.hdr.client_id; ++ hif->client_ctrl = (hif_hdr.hdr.client_ctrl1 << 16) | hif_hdr.hdr.client_ctrl; ++ flags = CL_DESC_FIRST; ++ ++// printk(KERN_INFO "start of packet: id %d, q %d, len %d, flags %x %x\n", hif->client_id, hif->qno, len, local_desc.ctrl, hif->client_ctrl); ++ } ++ else { ++// printk(KERN_INFO "continuation: id %d, q %d, len %d, flags %x\n", hif->client_id, hif->qno, len, local_desc.ctrl); ++ flags = 0; ++ } ++ ++ if (local_desc.ctrl & BD_CTRL_LIFM) ++ flags |= CL_DESC_LAST; ++#endif ++ /* Check for valid client id and still registered */ ++ if ((hif->client_id >= HIF_CLIENTS_MAX) || !(test_bit(hif->client_id, &hif->shm->gClient_status[0]))) { ++ if (printk_ratelimit()) ++ printk(KERN_ERR "%s: packet with invalid client id %d qNo %d\n", __func__, hif->client_id, hif->qno); ++ ++#if defined(CONFIG_PLATFORM_PCI) ++ free_buf = local_desc.data; ++#else ++ free_buf = pkt_hdr; ++#endif ++ goto pkt_drop; ++ } ++ ++ /* Check to valid queue number */ ++ if (hif->client[hif->client_id].rx_qn <= hif->qno) { ++ printk(KERN_INFO "%s: packet with invalid queue: %d\n", __func__, hif->qno); ++ hif->qno = 0; ++ } ++ ++#if defined(CONFIG_PLATFORM_PCI) ++ free_buf = client_put_rxpacket(&hif->client[hif->client_id].rx_q[hif->qno], ++ (void *)PFE_PCI_TO_HOST(desc->data), len, flags, hif->client_ctrl, &buf_zize); ++#else ++ free_buf = client_put_rxpacket(&hif->client[hif->client_id].rx_q[hif->qno], ++ (void *)pkt_hdr, len, flags, hif->client_ctrl, &buf_size); ++#endif ++ ++ hif_lib_indicate_client(hif->client_id, EVENT_RX_PKT_IND, hif->qno); ++ ++ if (unlikely(!free_buf)) { ++#ifdef HIF_NAPI_STATS ++ hif->napi_counters[NAPI_CLIENT_FULL_COUNT]++; ++#endif ++ /* If we want to keep in polling mode to retry later, we need to tell napi that we consumed ++ the full budget or we will hit a livelock scenario. The core code keeps this napi instance ++ at the head of the list and none of the other instances get to run */ ++ rx_processed = budget; ++ ++ if (flags & CL_DESC_FIRST) ++ hif->started = 0; ++ ++ break; ++ } ++ ++ pkt_drop: ++#if defined(CONFIG_PLATFORM_PCI) ++ desc->data = (u32)free_buf; ++#else ++ /*Fill free buffer in the descriptor */ ++ hif->rx_buf_addr[rtc] = free_buf; ++ hif->rx_buf_len[rtc] = min(pfe_pkt_size, buf_size); ++ desc->data = DDR_PHYS_TO_PFE((u32)DMA_MAP_SINGLE(hif->dev, free_buf, hif->rx_buf_len[rtc], DMA_FROM_DEVICE)); ++ //printk("#%p %p %d\n", desc->data, hif->rx_buf_addr[rtc], hif->rx_buf_len[rtc]); ++#endif ++ wmb(); ++ desc->ctrl = BD_CTRL_PKT_INT_EN | BD_CTRL_LIFM | BD_CTRL_DIR | ++ BD_CTRL_DESC_EN | BD_BUF_LEN(hif->rx_buf_len[rtc]); ++ ++ inc_hif_rxidx(rtc); ++ ++ if (local_desc.ctrl & BD_CTRL_LIFM) { ++ if (!(hif->client_ctrl & HIF_CTRL_RX_CONTINUED)) { ++ rx_processed++; ++ ++#ifdef HIF_NAPI_STATS ++ hif->napi_counters[NAPI_PACKET_COUNT]++; ++#endif ++ } ++ hif->started = 0; ++ } ++ } ++ ++ hif->RxtocleanIndex = rtc; ++ spin_unlock_bh(&hif->lock); ++ ++ /* we made some progress, re-start rx dma in case it stopped */ ++ hif_rx_dma_start(); ++ ++ return rx_processed; ++} ++ ++ ++/* client_ack_txpacket- ++ * This function ack the Tx packet in the give client Tx queue by resetting ++ * ownership bit in the descriptor. ++ */ ++static int client_ack_txpacket(struct pfe_hif *hif, unsigned int client_id, unsigned int q_no) ++{ ++ struct hif_tx_queue *queue = &hif->client[client_id].tx_q[q_no]; ++ struct tx_queue_desc *desc = queue->base + queue->ack_idx; ++ ++ if (desc->ctrl & CL_DESC_OWN) { ++ /*TODO Do we need to match the pkt address also? */ ++ desc->ctrl &= ~CL_DESC_OWN; ++ inc_cl_idx(queue->ack_idx); ++ ++ return 0; ++ } ++ else { ++ /*This should not happen */ ++ printk(KERN_ERR "%s: %d %d %d %d %d %p %d\n", __func__, hif->Txtosend, hif->Txtoclean, hif->TxAvail, client_id, q_no, queue, queue->ack_idx); ++ BUG(); ++ return 1; ++ } ++} ++ ++void __hif_tx_done_process(struct pfe_hif *hif, int count) ++{ ++ struct hif_desc *desc; ++ struct hif_desc_sw *desc_sw; ++ int ttc, tx_avl; ++ ++ ttc = hif->Txtoclean; ++ tx_avl = hif->TxAvail; ++ ++ while ((tx_avl < hif->TxRingSize) && count--) { ++ desc = hif->TxBase + ttc; ++ ++ if (desc->ctrl & BD_CTRL_DESC_EN) ++ break; ++ ++ desc_sw = &hif->tx_sw_queue[ttc]; ++ ++ if (desc_sw->data) { ++#if !defined(CONFIG_PLATFORM_PCI) ++ //dmap_unmap_single(hif->dev, desc_sw->data, desc_sw->len, DMA_TO_DEVICE); ++ DMA_UNMAP_SINGLE(hif->dev, desc_sw->data, desc_sw->len, DMA_TO_DEVICE); ++#endif ++ } ++ client_ack_txpacket(hif, desc_sw->client_id, desc_sw->q_no); ++ ++ inc_hif_txidx(ttc); ++ tx_avl++; ++ } ++ ++ hif->Txtoclean = ttc; ++ hif->TxAvail = tx_avl; ++} ++ ++ ++/* __hif_xmit_pkt - ++ * This function puts one packet in the HIF Tx queue ++ */ ++void __hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int q_no, void *data, u32 len, unsigned int flags) ++{ ++ struct hif_desc *desc; ++ struct hif_desc_sw *desc_sw; ++ ++#if defined(CONFIG_PLATFORM_EMULATION) ++ { ++ struct hif_queue *queue = &hif->client[client_id].rx_q[0]; ++ struct queue_desc *qdesc = queue->base + queue->write_idx; ++ void *buf; ++ ++ printk("%s: packet loop backed client_id:%d qno:%d data : %p len:%d\n", __func__, client_id, q_no, data, len); ++#if 1 ++ if (qdesc->ctrl & CL_DESC_OWN) { ++ buf = (void *)qdesc->data; ++ memcpy(buf, data, len); ++ wmb(); ++ qdesc->ctrl = CL_DESC_BUF_LEN(len); ++ inc_cl_idx(queue->write_idx); ++ printk("%s: packet loop backed..\n", __func__); ++ hif_lib_indicate_client(client_id, EVENT_RX_PKT_IND, q_no); ++ client_ack_txpacket(&hif->client[client_id].tx_q[q_no]); ++ } ++#endif ++ } ++ ++#else ++ desc = hif->TxBase + hif->Txtosend; ++ desc_sw = &hif->tx_sw_queue[hif->Txtosend]; ++ ++ desc_sw->len = len; ++ desc_sw->client_id = client_id; ++ desc_sw->q_no = q_no; ++ desc_sw->flags = flags; ++ ++#if !defined(CONFIG_PLATFORM_PCI) ++ if (flags & HIF_DONT_DMA_MAP) { ++ desc_sw->data = 0; ++ desc->data = (u32)DDR_PHYS_TO_PFE(data); ++ } else { ++ desc_sw->data = DMA_MAP_SINGLE(hif->dev, data, len, DMA_TO_DEVICE); ++ desc->data = (u32)DDR_PHYS_TO_PFE(desc_sw->data); ++ } ++#else ++#define ALIGN32(x) ((x) & ~0x3) ++ memcpy(PFE_PCI_TO_HOST(desc->data), data, ALIGN32(len+0x3)); ++#endif ++ ++ inc_hif_txidx(hif->Txtosend); ++ hif->TxAvail--; ++ ++ /* For TSO we skip actual TX until the last descriptor */ ++ /* This reduce the number of required wmb() */ ++ if ((flags & HIF_TSO) && (!((flags & HIF_DATA_VALID) && (flags & HIF_LAST_BUFFER)))) ++ goto skip_tx; ++ ++ wmb(); ++ ++ do { ++ desc_sw = &hif->tx_sw_queue[hif->Txtoflush]; ++ desc = hif->TxBase + hif->Txtoflush; ++ ++ if (desc_sw->flags & HIF_LAST_BUFFER) { ++ if ((desc_sw->client_id < PFE_CL_VWD0) || (desc_sw->client_id > (PFE_CL_VWD0 + MAX_VAP_SUPPORT))) ++ desc->ctrl = BD_CTRL_LIFM | BD_CTRL_BRFETCH_DISABLE | ++ BD_CTRL_RTFETCH_DISABLE | BD_CTRL_PARSE_DISABLE | ++ BD_CTRL_DESC_EN | BD_BUF_LEN(desc_sw->len); ++ else { ++ ++ desc->ctrl = BD_CTRL_LIFM | BD_CTRL_DESC_EN | BD_BUF_LEN(desc_sw->len); ++ } ++ } ++ else ++ desc->ctrl = BD_CTRL_DESC_EN | BD_BUF_LEN(desc_sw->len); ++ ++ inc_hif_txidx(hif->Txtoflush); ++ } ++ while (hif->Txtoflush != hif->Txtosend); ++ ++skip_tx: ++ return; ++ ++#endif ++} ++ ++ ++int hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int q_no, void *data, unsigned int len) ++{ ++ int rc = 0; ++ ++ spin_lock_bh(&hif->tx_lock); ++ ++ if (!hif->TxAvail) ++ rc = 1; ++ else { ++ __hif_xmit_pkt(hif, client_id, q_no, data, len, HIF_FIRST_BUFFER | HIF_LAST_BUFFER); ++ hif_tx_dma_start(); ++ } ++ if (hif->TxAvail < (hif->TxRingSize >> 1)) ++ __hif_tx_done_process(hif, TX_FREE_MAX_COUNT); ++ ++ spin_unlock_bh(&hif->tx_lock); ++ ++ return rc; ++} ++ ++/* hif_isr- ++ * This ISR routine processes Rx/Tx done interrupts from the HIF hardware block ++ */ ++static irqreturn_t hif_isr(int irq, void *dev_id) ++{ ++ struct pfe_hif *hif = (struct pfe_hif *) dev_id; ++ int int_status; ++ ++ /*Read hif interrupt source register */ ++ int_status = readl_relaxed(HIF_INT_SRC); ++ ++ if ((int_status & HIF_INT) == 0) ++ return(IRQ_NONE); ++ ++ int_status &= ~(HIF_INT); ++ ++ if (int_status & HIF_RXPKT_INT) { ++ int_status &= ~(HIF_RXPKT_INT); ++ ++ /* Disable interrupts */ ++ writel_relaxed(0, HIF_INT_ENABLE); ++ ++ napi_first_batch = 1; ++ ++ if (napi_schedule_prep(&hif->napi)) ++ { ++#ifdef HIF_NAPI_STATS ++ hif->napi_counters[NAPI_SCHED_COUNT]++; ++#endif ++ __napi_schedule(&hif->napi); ++ } ++ } ++ ++ if (int_status) { ++ printk(KERN_INFO "%s : Invalid interrupt : %d\n", __func__, int_status); ++ writel(int_status, HIF_INT_SRC); ++ } ++ ++ return IRQ_HANDLED; ++} ++ ++ ++void hif_process_client_req(struct pfe_hif *hif, int req, int data1, int data2) ++{ ++ unsigned int client_id = data1; ++ ++ if (client_id >= HIF_CLIENTS_MAX) ++ { ++ printk(KERN_ERR "%s: client id %d out of bounds\n", __func__, client_id); ++ return; ++ } ++ ++ switch (req) { ++ case REQUEST_CL_REGISTER: ++ /* Request for register a client */ ++ printk(KERN_INFO "%s: register client_id %d\n", __func__, client_id); ++ pfe_hif_client_register(hif, client_id, (struct hif_client_shm *)&hif->shm->client[client_id]); ++ break; ++ ++ case REQUEST_CL_UNREGISTER: ++ printk(KERN_INFO "%s: unregister client_id %d\n", __func__, client_id); ++ ++ /* Request for unregister a client */ ++ pfe_hif_client_unregister(hif, client_id); ++ ++ break; ++ ++ default: ++ printk(KERN_ERR "%s: unsupported request %d\n", __func__, req); ++ break; ++ } ++ ++ /*TODO check for TMU queue resume request */ ++ ++ /*Process client Tx queues ++ * Currently we don't have checking for tx pending*/ ++} ++ ++/** pfe_hif_rx_poll ++ * This function is NAPI poll function to process HIF Rx queue. ++ */ ++static int pfe_hif_rx_poll(struct napi_struct *napi, int budget) ++{ ++ struct pfe_hif *hif = container_of(napi, struct pfe_hif, napi); ++ int work_done; ++ ++#ifdef HIF_NAPI_STATS ++ hif->napi_counters[NAPI_POLL_COUNT]++; ++#endif ++ ++ work_done = pfe_hif_rx_process(hif, budget); ++ ++ if (work_done < budget) ++ { ++ napi_complete(napi); ++ writel_relaxed(HIF_INT_MASK, HIF_INT_ENABLE); ++ } ++#ifdef HIF_NAPI_STATS ++ else ++ hif->napi_counters[NAPI_FULL_BUDGET_COUNT]++; ++#endif ++ ++ return work_done; ++} ++ ++/* pfe_hif_init ++ * This function initializes the baseaddresses and irq, etc. ++ */ ++int pfe_hif_init(struct pfe *pfe) ++{ ++ struct pfe_hif *hif = &pfe->hif; ++ int err; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ hif->dev = pfe->dev; ++ hif->irq = pfe->hif_irq; ++ ++ if ((err = pfe_hif_alloc_descr(hif))) { ++ goto err0; ++ } ++ ++ if (pfe_hif_init_buffers(hif)) { ++ printk(KERN_ERR "%s: Could not initialize buffer descriptors\n", __func__); ++ err = -ENOMEM; ++ goto err1; ++ } ++ ++ /* Initilize NAPI for Rx processing */ ++ init_dummy_netdev(&hif->dummy_dev); ++ netif_napi_add(&hif->dummy_dev, &hif->napi, pfe_hif_rx_poll, HIF_RX_POLL_WEIGHT); ++ napi_enable(&hif->napi); ++ ++ spin_lock_init(&hif->tx_lock); ++ spin_lock_init(&hif->lock); ++ ++ hif_init(); ++ hif_rx_enable(); ++ hif_tx_enable(); ++ ++ /* Disable tx done interrupt */ ++ writel(HIF_INT_MASK, HIF_INT_ENABLE); ++ ++ gpi_enable(HGPI_BASE_ADDR); ++ ++#ifdef __KERNEL__ ++ err = request_irq(hif->irq, hif_isr, 0, "pfe_hif", hif); ++ if (err) { ++ printk(KERN_ERR "%s: failed to get the hif IRQ = %d\n", __func__, hif->irq); ++ goto err1; ++ } ++#else ++ /*TODO register interrupts */ ++#endif ++ ++ return 0; ++err1: ++ pfe_hif_free_descr(hif); ++err0: ++ return err; ++} ++ ++/* pfe_hif_exit- ++ */ ++void pfe_hif_exit(struct pfe *pfe) ++{ ++ struct pfe_hif *hif = &pfe->hif; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ spin_lock_bh(&hif->lock); ++ hif->shm->gClient_status[0] = 0; ++ hif->shm->gClient_status[1] = 0; /* Make sure all clients are disabled */ ++ ++ spin_unlock_bh(&hif->lock); ++ ++ /*Disable Rx/Tx */ ++ gpi_disable(HGPI_BASE_ADDR); ++ hif_rx_disable(); ++ hif_tx_disable(); ++ ++ napi_disable(&hif->napi); ++ netif_napi_del(&hif->napi); ++ ++#ifdef __KERNEL__ ++ free_irq(hif->irq, hif); ++#endif ++ pfe_hif_release_buffers(hif); ++ pfe_hif_free_descr(hif); ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hif.h +@@ -0,0 +1,322 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_HIF_H_ ++#define _PFE_HIF_H_ ++ ++#include <linux/netdevice.h> ++ ++#define HIF_NAPI_STATS ++ ++#define HIF_CLIENT_QUEUES_MAX 16 ++#define HIF_RX_POLL_WEIGHT 64 ++ ++#define HIF_RX_PKT_MIN_SIZE 0x800 /* 2KB */ ++#define HIF_RX_PKT_MIN_SIZE_MASK ~(HIF_RX_PKT_MIN_SIZE - 1) ++#define ROUND_MIN_RX_SIZE(_sz) ((_sz + (HIF_RX_PKT_MIN_SIZE - 1)) & HIF_RX_PKT_MIN_SIZE_MASK) ++#define PRESENT_OFST_IN_PAGE(_buf) (((unsigned long int)_buf & (PAGE_SIZE - 1)) & HIF_RX_PKT_MIN_SIZE_MASK) ++ ++enum { ++ NAPI_SCHED_COUNT = 0, ++ NAPI_POLL_COUNT, ++ NAPI_PACKET_COUNT, ++ NAPI_DESC_COUNT, ++ NAPI_FULL_BUDGET_COUNT, ++ NAPI_CLIENT_FULL_COUNT, ++ NAPI_MAX_COUNT ++}; ++ ++ ++/* XXX HIF_TX_DESC_NT value should be always greter than 4, ++ * Otherwise HIF_TX_POLL_MARK will become zero. ++ */ ++#if defined(CONFIG_PLATFORM_PCI) ++#define HIF_RX_DESC_NT 4 ++#define HIF_TX_DESC_NT 4 ++#else ++#if defined(CONFIG_COMCERTO_64K_PAGES) ++#define HIF_RX_DESC_NT 64 ++#else ++#define HIF_RX_DESC_NT 256 ++#endif ++#define HIF_TX_DESC_NT 2048 ++#endif ++ ++#define HIF_FIRST_BUFFER (1 << 0) ++#define HIF_LAST_BUFFER (1 << 1) ++#define HIF_DONT_DMA_MAP (1 << 2) //TODO merge it with TSO ++#define HIF_DATA_VALID (1 << 3) ++#define HIF_TSO (1 << 4) ++ ++#define MAX_VAP_SUPPORT 3 ++#define MAX_WIFI_VAPS MAX_VAP_SUPPORT ++ ++enum { ++ PFE_CL_GEM0 = 0, ++ PFE_CL_GEM1, ++ PFE_CL_GEM2, ++ PFE_CL_VWD0, ++ PFE_CL_VWD_LAST = PFE_CL_VWD0 + MAX_VAP_SUPPORT, ++ PFE_CL_PCAP0, ++ HIF_CLIENTS_MAX ++}; ++ ++/*structure to store client queue info */ ++struct hif_rx_queue { ++ struct rx_queue_desc *base; ++ u32 size; ++ u32 write_idx; ++}; ++ ++struct hif_tx_queue { ++ struct tx_queue_desc *base; ++ u32 size; ++ u32 ack_idx; ++}; ++ ++/*Structure to store the client info */ ++struct hif_client { ++ int rx_qn; ++ struct hif_rx_queue rx_q[HIF_CLIENT_QUEUES_MAX]; ++ int tx_qn; ++ struct hif_tx_queue tx_q[HIF_CLIENT_QUEUES_MAX]; ++}; ++ ++/*HIF hardware buffer descriptor */ ++struct hif_desc { ++ volatile u32 ctrl; ++ volatile u32 status; ++ volatile u32 data; ++ volatile u32 next; ++}; ++ ++struct __hif_desc { ++ u32 ctrl; ++ u32 status; ++ u32 data; ++}; ++ ++struct hif_desc_sw { ++ dma_addr_t data; ++ u16 len; ++ u8 client_id; ++ u8 q_no; ++ u16 flags; ++}; ++ ++struct hif_hdr { ++ u8 client_id; ++ u8 qNo; ++ u16 client_ctrl; ++ u16 client_ctrl1; ++}; ++ ++struct __hif_hdr { ++ union { ++ struct hif_hdr hdr; ++ u32 word[2]; ++ }; ++}; ++ ++struct hif_lro_hdr { ++ u16 data_offset; ++ u16 mss; ++}; ++ ++struct hif_ipsec_hdr { ++ u16 sa_handle[2]; ++}__attribute__((packed)); ++ ++#define MAX_TSO_BUF_DESCS 5 ++struct hif_tso_buf_desc { ++ u32 addr; ++ u32 ctrl; ++#define TSO_CTRL_LAST_BUFFER (1 << 31) ++}; ++ ++struct hif_tso_hdr { ++ struct hif_hdr pkt_hdr; ++ u16 ip_off; ++ u16 ip_id; ++ u16 ip_len; ++ u16 tcp_off; ++ u32 tcp_seq; ++} __attribute__((packed)); ++ ++struct hif_tso_hdr_nocpy { ++ struct hif_tso_hdr tso_hdr; ++ struct hif_tso_buf_desc bdesc[MAX_TSO_BUF_DESCS]; ++} __attribute__((packed)); ++ ++struct hif_pcap_hdr { ++ u8 ifindex; ++ u8 unused; ++ u16 seqno; ++ u32 timestamp; ++}__attribute__((packed)); ++ ++/* HIF_CTRL_TX... defines */ ++#define HIF_CTRL_TX_TSO_NOCPY (1 << 8) ++#define HIF_CTRL_TX_IPSEC_OUT (1 << 7) ++#define HIF_CTRL_TX_OWN_MAC (1 << 6) ++#define HIF_CTRL_TX_TSO_END (1 << 5) ++#define HIF_CTRL_TX_TSO6 (1 << 4) ++#define HIF_CTRL_TX_TSO (1 << 3) ++#define HIF_CTRL_TX_CHECKSUM (1 << 2) ++#define HIF_CTRL_TX_CSUM_VALIDATE (1 << 1) ++#define HIF_CTRL_TX_WIFI (1 << 0) ++ ++/* HIF_CTRL_RX... defines */ ++#define HIF_CTRL_RX_OFFSET_OFST (24) ++#define HIF_CTRL_RX_PE_ID_OFST (16) ++#define HIF_CTRL_RX_IPSEC_IN (1 << 4) ++#define HIF_CTRL_RX_WIFI_EXPT (1 << 3) ++#define HIF_CTRL_RX_CHECKSUMMED (1 << 2) ++#define HIF_CTRL_RX_CONTINUED (1 << 1) ++#define HIF_CTRL_RX_WIFI_HEADROOM (1 << 0) ++ ++#define HIF_CTRL_VAPID_OFST (8) ++ ++struct pfe_hif { ++ /* To store registered clients in hif layer */ ++ struct hif_client client[HIF_CLIENTS_MAX]; ++ struct hif_shm *shm; ++ int irq; ++ ++ void *descr_baseaddr_v; ++ unsigned long descr_baseaddr_p; ++ ++ struct hif_desc *RxBase; ++ u32 RxRingSize; ++ u32 RxtocleanIndex; ++ void *rx_buf_addr[HIF_RX_DESC_NT]; ++ int rx_buf_len[HIF_RX_DESC_NT]; ++ unsigned int qno; ++ unsigned int client_id; ++ unsigned int client_ctrl; ++ unsigned int started; ++ ++ struct hif_desc *TxBase; ++ u32 TxRingSize; ++ u32 Txtosend; ++ u32 Txtoclean; ++ u32 TxAvail; ++ u32 Txtoflush; ++ struct hif_desc_sw tx_sw_queue[HIF_TX_DESC_NT]; ++ struct hif_tso_hdr_nocpy *tso_hdr_v; ++ dma_addr_t tso_hdr_p; ++ ++ spinlock_t tx_lock; ++ spinlock_t lock; ++ struct net_device dummy_dev; ++ struct napi_struct napi; ++ struct device *dev; ++ ++#ifdef CONFIG_HOTPLUG_CPU ++ struct notifier_block cpu_notify; ++#endif ++ ++#ifdef HIF_NAPI_STATS ++ unsigned int napi_counters[NAPI_MAX_COUNT]; ++#endif ++}; ++ ++void __hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int q_no, void *data, u32 len, unsigned int flags); ++int hif_xmit_pkt(struct pfe_hif *hif, unsigned int client_id, unsigned int q_no, void *data, unsigned int len); ++void __hif_tx_done_process(struct pfe_hif *hif, int count); ++void hif_process_client_req(struct pfe_hif *hif, int req, int data1, int data2); ++int pfe_hif_init(struct pfe *pfe); ++void pfe_hif_exit(struct pfe *pfe); ++ ++static inline void hif_tx_done_process(struct pfe_hif *hif, int count) ++{ ++ spin_lock_bh(&hif->tx_lock); ++ __hif_tx_done_process(hif, count); ++ spin_unlock_bh(&hif->tx_lock); ++} ++ ++static inline void hif_tx_lock(struct pfe_hif *hif) ++{ ++ spin_lock_bh(&hif->tx_lock); ++} ++ ++static inline void hif_tx_unlock(struct pfe_hif *hif) ++{ ++ spin_unlock_bh(&hif->tx_lock); ++} ++ ++static inline int __hif_tx_avail(struct pfe_hif *hif) ++{ ++ return hif->TxAvail; ++} ++ ++#if defined(CONFIG_PLATFORM_C2000) ++static inline void __memcpy8(void *dst, void *src) ++{ ++ asm volatile ( "ldm %1, {r9, r10}\n\t" ++ "stm %0, {r9, r10}\n\t" ++ : ++ : "r" (dst), "r" (src) ++ : "r9", "r10", "memory" ++ ); ++} ++ ++static inline void __memcpy12(void *dst, void *src) ++{ ++ asm volatile ( "ldm %1, {r8, r9, r10}\n\t" ++ "stm %0, {r8, r9, r10}\n\t" ++ : ++ : "r" (dst), "r" (src) ++ : "r8", "r9", "r10", "memory" ++ ); ++} ++ ++static inline void __memcpy16(void *dst, void *src) ++{ ++ asm volatile ( "ldm %1, {r7, r8, r9, r10}\n\t" ++ "stm %0, {r7, r8, r9, r10}\n\t" ++ : ++ : "r"(dst), "r"(src) ++ : "r7", "r8", "r9", "r10", "memory" ++ ); ++} ++ ++#define HIF_MEMCPY_BURSTSIZE 32 /*__memcpy copy 32byte in a burst*/ ++static inline void __memcpy(void *dst, void *src, unsigned int len) ++{ ++ void *end = src + len; ++ ++ dst = (void *)((unsigned long)dst & ~0x3); ++ src = (void *)((unsigned long)src & ~0x3); ++ ++ while (src < end) { ++ asm volatile ( "ldm %1!, {r3, r4, r5, r6, r7, r8, r9, r10}\n\t" ++ "stm %0!, {r3, r4, r5, r6, r7, r8, r9, r10}\n\t" ++ : "+r"(dst), "+r"(src) ++ : ++ : "r3", "r4", "r5", "r6", "r7", "r8", "r9", "r10", "memory" ++ ); ++ } ++} ++#else ++#define __memcpy8(dst, src) memcpy(dst, src, 8) ++#define __memcpy12(dst, src) memcpy(dst, src, 12) ++#define __memcpy(dst, src, len) memcpy(dst, src, len) ++#endif ++#endif /* _PFE_HIF_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hif_lib.c +@@ -0,0 +1,658 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#include <linux/version.h> ++#if LINUX_VERSION_CODE < KERNEL_VERSION(3,10,0) ++#include <asm/system.h> ++#endif ++#include <linux/kernel.h> ++#include <linux/slab.h> ++#include <linux/interrupt.h> ++#include <linux/workqueue.h> ++#include <linux/dma-mapping.h> ++#include <linux/dmapool.h> ++#include <linux/sched.h> ++#include <linux/skbuff.h> ++#include <linux/moduleparam.h> ++#include <linux/cpu.h> ++#if LINUX_VERSION_CODE >= KERNEL_VERSION(3,10,0) ++//#include <asm/system_info.h> ++#endif ++ ++#include "pfe_mod.h" ++#include "pfe_hif.h" ++#include "pfe_hif_lib.h" ++#include "pfe_ctrl_hal.h" ++ ++ ++unsigned int lro_mode = 0; ++unsigned int page_mode = 0; ++unsigned int tx_qos = 0; ++unsigned int pfe_pkt_size; ++unsigned int pfe_pkt_headroom; ++unsigned int emac_txq_cnt; ++ ++/** @pfe_hal_lib.c. ++ * Common functions used by HIF client drivers ++ */ ++ ++/*HIF shared memory Global variable */ ++struct hif_shm ghif_shm; ++ ++/* TMU tx transmitted packets counter, 1 per TMU */ ++unsigned int TMU_DMEM_SH(tx_trans)[EMAC_TXQ_CNT]; ++ ++/* Cleanup the HIF shared memory, release HIF rx_buffer_pool. ++ * This function should be called after pfe_hif_exit ++ * ++ * @param[in] hif_shm Shared memory address location in DDR ++ */ ++static void pfe_hif_shm_clean(struct hif_shm *hif_shm) ++{ ++ int i; ++ void *pkt; ++ ++ for (i = 0; i < hif_shm->rx_buf_pool_cnt; i++) { ++ pkt = hif_shm->rx_buf_pool[i]; ++ if (pkt) { ++ hif_shm->rx_buf_pool[i] = NULL; ++ pkt -= pfe_pkt_headroom; ++ ++ if (page_mode) { ++ put_page(virt_to_page(pkt)); ++ } else ++ kfree(pkt); ++ } ++ } ++} ++ ++/* Initialize shared memory used between HIF driver and clients, ++ * allocate rx_buffer_pool required for HIF Rx descriptors. ++ * This function should be called before initializing HIF driver. ++ * ++ * @param[in] hif_shm Shared memory address location in DDR ++ * @rerurn 0 - on succes, <0 on fail to initialize ++ */ ++static int pfe_hif_shm_init(struct hif_shm *hif_shm) ++{ ++ int i; ++ void *pkt; ++ ++ memset(hif_shm, 0, sizeof(struct hif_shm)); ++ hif_shm->rx_buf_pool_cnt = HIF_RX_DESC_NT; ++ ++ for (i = 0; i < hif_shm->rx_buf_pool_cnt; i++) { ++ if (page_mode) { ++ pkt = (void *)__get_free_page(GFP_KERNEL | GFP_DMA_PFE); ++ } else ++ pkt = kmalloc(PFE_BUF_SIZE, GFP_KERNEL | GFP_DMA_PFE); ++ ++ if (pkt) ++ hif_shm->rx_buf_pool[i] = pkt + pfe_pkt_headroom; ++ else ++ goto err0; ++ } ++ ++ return 0; ++ ++err0: ++ printk(KERN_ERR "%s Low memory\n", __func__); ++ pfe_hif_shm_clean(hif_shm); ++ return -ENOMEM; ++} ++ ++/*This function sends indication to HIF driver ++ * ++ * @param[in] hif hif context ++ **/ ++static void hif_lib_indicate_hif(struct pfe_hif *hif, int req, int data1, int data2) ++{ ++ //TODO : If we separate HIF and HIF LIB, then send req and data through shared memory. ++ ++ hif_process_client_req(hif, req, data1, data2); ++} ++ ++void hif_lib_indicate_client(int client_id, int event_type, int qno) ++{ ++ struct hif_client_s *client = pfe->hif_client[client_id]; ++ ++ /* ++ * TODO : Right now, all events are queue number based. So we are masking events per queue ++ * basis. Later if we add any events those do not depend on queue number, then we may ++ * may need may need to add masking per event. ++ */ ++ if (!client || (event_type >= HIF_EVENT_MAX) || (qno >= HIF_CLIENT_QUEUES_MAX) ) ++ return; ++ ++ if (!test_and_set_bit(qno, &client->queue_mask[event_type])) { ++ client->event_handler(client->priv, event_type, qno); ++ } ++ ++} ++ ++ ++/*This function releases Rx queue descriptors memory and pre-filled buffers ++ * ++ * @param[in] client hif_client context ++ */ ++static void hif_lib_client_release_rx_buffers(struct hif_client_s *client) ++{ ++ struct rx_queue_desc *desc; ++ int qno, ii; ++ void *buf; ++ ++ for (qno = 0; qno < client->rx_qn; qno++) { ++ desc = client->rx_q[qno].base; ++ ++ for (ii = 0; ii < client->rx_q[qno].size; ii++) { ++ buf = (void *)desc->data; ++ if (buf) { ++ buf -= pfe_pkt_headroom; ++ ++ if (page_mode) ++ free_page((unsigned long)buf); ++ else ++ kfree(buf); ++ ++ desc->ctrl = 0; ++ } ++ ++ desc++; ++ } ++ } ++ ++ kfree(client->rx_qbase); ++} ++ ++ ++/*This function allocates memory for the rxq descriptors and pre-fill rx queues ++ * with buffers. ++ * @param[in] client client context ++ * @param[in] q_size size of the rxQ, all queues are of same size ++ */ ++static int hif_lib_client_init_rx_buffers(struct hif_client_s *client, int q_size) ++{ ++ struct rx_queue_desc *desc; ++ struct hif_client_rx_queue *queue; ++ int ii, qno; ++ ++ /*Allocate memory for the client queues */ ++ client->rx_qbase = kzalloc(client->rx_qn * q_size * sizeof(struct rx_queue_desc), GFP_KERNEL); ++ if (!client->rx_qbase){ ++ goto err; ++ } ++ ++ for (qno = 0; qno < client->rx_qn; qno++) { ++ queue = &client->rx_q[qno]; ++ ++ queue->base = client->rx_qbase + qno * q_size * sizeof(struct rx_queue_desc); ++ queue->size = q_size; ++ queue->read_idx = 0; ++ queue->write_idx = 0; ++ ++ dbg_print_info("rx queue: %d, base: %p, size: %d \n", qno, queue->base, queue->size); ++ } ++ ++ for (qno = 0; qno < client->rx_qn; qno++) { ++ queue = &client->rx_q[qno]; ++ desc = queue->base; ++ ++ for (ii = 0; ii < queue->size; ii++) { ++ desc->ctrl = CL_DESC_BUF_LEN(pfe_pkt_size) | CL_DESC_OWN; ++ desc++; ++ } ++ } ++ ++ return 0; ++ ++err: ++ return 1; ++} ++ ++#define inc_cl_idx(idxname) idxname = (idxname+1) & (queue->size-1) ++ ++static void hif_lib_client_cleanup_tx_queue(struct hif_client_tx_queue *queue) ++{ ++ dbg_print_info( "%s\n", __func__); ++ ++ /* Check if there are any pending packets. Client must flush the tx queues ++ before unregistering, by calling by calling hif_lib_tx_get_next_complete() */ ++ /* Hif no longer calls since we are no longer registered */ ++ ++ if (queue->tx_pending) ++ printk(KERN_ERR "%s: pending transmit packets\n", __func__); ++} ++ ++static void hif_lib_client_release_tx_buffers(struct hif_client_s *client) ++{ ++ int qno; ++ ++ dbg_print_info("%s\n", __func__); ++ ++ for (qno = 0; qno < client->tx_qn; qno++) { ++ hif_lib_client_cleanup_tx_queue(&client->tx_q[qno]); ++ } ++ ++ kfree(client->tx_qbase); ++} ++ ++static int hif_lib_client_init_tx_buffers(struct hif_client_s *client, int q_size) ++{ ++ struct hif_client_tx_queue *queue; ++ int qno; ++ ++ client->tx_qbase = kzalloc(client->tx_qn * q_size * sizeof(struct tx_queue_desc), GFP_KERNEL); ++ if (!client->tx_qbase) { ++ return 1; ++ } ++ ++ for (qno = 0; qno < client->tx_qn; qno++) { ++ queue = &client->tx_q[qno]; ++ ++ queue->base = client->tx_qbase + qno * q_size * sizeof(struct tx_queue_desc); ++ queue->size = q_size; ++ queue->read_idx = 0; ++ queue->write_idx = 0; ++ queue->tx_pending = 0; ++ queue->nocpy_flag = 0; ++ queue->prev_tmu_tx_pkts = 0; ++ queue->done_tmu_tx_pkts = 0; ++ ++ dbg_print_info("tx queue: %d, base: %p, size: %d \n", qno, queue->base, queue->size); ++ } ++ ++ return 0; ++} ++ ++static int hif_lib_event_dummy( void *priv, int event_type, int qno) ++{ ++ return 0; ++} ++ ++int hif_lib_client_register(struct hif_client_s *client) ++{ ++ struct hif_shm *hif_shm; ++ struct hif_client_shm *client_shm; ++ int err, i; ++// int loop_cnt = 0; ++ ++ dbg_print_info("%s\n", __func__); ++ ++ /*Allocate memory before spin_lock*/ ++ if (hif_lib_client_init_rx_buffers(client, client->rx_qsize)) { ++ err = -ENOMEM; ++ goto err_rx; ++ } ++ ++ if (hif_lib_client_init_tx_buffers(client, client->tx_qsize)) { ++ err = -ENOMEM; ++ goto err_tx; ++ } ++ ++ spin_lock_bh(&pfe->hif.lock); ++ if (!(client->pfe) || (client->id >= HIF_CLIENTS_MAX) || (pfe->hif_client[client->id])) { ++ err = -EINVAL; ++ goto err; ++ } ++ ++ hif_shm = client->pfe->hif.shm; ++ ++ if (!client->event_handler) ++ client->event_handler = hif_lib_event_dummy; ++ ++ /*Initialize client specific shared memory */ ++ client_shm = (struct hif_client_shm *)&hif_shm->client[client->id]; ++ client_shm->rx_qbase = (unsigned long int)client->rx_qbase; ++ client_shm->rx_qsize = client->rx_qsize; ++ client_shm->tx_qbase = (unsigned long int)client->tx_qbase; ++ client_shm->tx_qsize = client->tx_qsize; ++ client_shm->ctrl = (client->tx_qn << CLIENT_CTRL_TX_Q_CNT_OFST) | (client->rx_qn << CLIENT_CTRL_RX_Q_CNT_OFST); ++// spin_lock_init(&client->rx_lock); ++ ++ for (i = 0; i < HIF_EVENT_MAX; i++) { ++ client->queue_mask[i] = 0; /* By default all events are unmasked */ ++ } ++ ++ /*Indicate to HIF driver*/ ++ hif_lib_indicate_hif(&pfe->hif, REQUEST_CL_REGISTER, client->id, 0); ++ ++ dbg_print_info("%s: client: %p, client_id: %d, tx_qsize: %d, rx_qsize: %d\n", ++ __func__, client, client->id, client->tx_qsize, client->rx_qsize); ++ ++ client->cpu_id = -1; ++ ++ pfe->hif_client[client->id] = client; ++ spin_unlock_bh(&pfe->hif.lock); ++ ++ return 0; ++ ++err: ++ spin_unlock_bh(&pfe->hif.lock); ++ hif_lib_client_release_tx_buffers(client); ++ ++err_tx: ++ hif_lib_client_release_rx_buffers(client); ++ ++err_rx: ++ return err; ++} ++ ++int hif_lib_client_unregister(struct hif_client_s *client) ++{ ++ struct pfe *pfe = client->pfe; ++ u32 client_id = client->id; ++ ++ printk(KERN_INFO "%s : client: %p, client_id: %d, txQ_depth: %d, rxQ_depth: %d\n", ++ __func__, client, client->id, client->tx_qsize, client->rx_qsize); ++ ++ ++ spin_lock_bh(&pfe->hif.lock); ++ hif_lib_indicate_hif(&pfe->hif, REQUEST_CL_UNREGISTER, client->id, 0); ++ ++ hif_lib_client_release_tx_buffers(client); ++ hif_lib_client_release_rx_buffers(client); ++ pfe->hif_client[client_id] = NULL; ++ spin_unlock_bh(&pfe->hif.lock); ++ ++ return 0; ++} ++ ++int hif_lib_event_handler_start(struct hif_client_s *client, int event, int qno) ++{ ++ struct hif_client_rx_queue *queue = &client->rx_q[qno]; ++ struct rx_queue_desc *desc = queue->base + queue->read_idx; ++ ++ if ((event >= HIF_EVENT_MAX) || ( qno >= HIF_CLIENT_QUEUES_MAX)) { ++ dbg_print_info("%s: Unsupported event : %d queue number : %d\n", __func__, event, qno); ++ return -1; ++ } ++ ++ test_and_clear_bit(qno, &client->queue_mask[event]); ++ ++ switch (event) { ++ case EVENT_RX_PKT_IND: ++ if (!(desc->ctrl & CL_DESC_OWN)) ++ hif_lib_indicate_client(client->id, EVENT_RX_PKT_IND, qno); ++ break; ++ ++ case EVENT_HIGH_RX_WM: ++ case EVENT_TXDONE_IND: ++ default: ++ break; ++ } ++ ++ return 0; ++} ++ ++ ++/*This function gets one packet from the specified client queue ++ * It also refill the rx buffer */ ++void *hif_lib_receive_pkt(struct hif_client_s *client, int qno, int *len, int *ofst, unsigned int *rx_ctrl, unsigned int *desc_ctrl, void **priv_data) ++{ ++ struct hif_client_rx_queue *queue = &client->rx_q[qno]; ++ struct rx_queue_desc *desc; ++ void *pkt = NULL; ++ ++ //printk(KERN_INFO "%s\n", __func__); ++#if defined(CONFIG_PLATFORM_EMULATION) ++ printk(KERN_INFO "%s:qno:%d cid:%d desc:%p rdidx:%d \n", ++ __func__, qno, client->id, desc, ++ queue->read_idx); ++#endif ++ ++ /* Following lock is to protect rx queue access from, hif_lib_event_handler_start. ++ * In general below lock is not required, because hif_lib_xmit_pkt and ++ * hif_lib_event_handler_start are called from napi poll and which is not ++ * re-entrant. But if some client use in different way this lock is required. ++ */ ++ //spin_lock_irqsave(&client->rx_lock, flags); ++ desc = queue->base + queue->read_idx; ++ if (!(desc->ctrl & CL_DESC_OWN)) { ++ pkt = desc->data - pfe_pkt_headroom; ++ ++ *rx_ctrl = desc->client_ctrl; ++ *desc_ctrl = desc->ctrl; ++ ++ if (desc->ctrl & CL_DESC_FIRST) { ++ u16 size = *rx_ctrl >> HIF_CTRL_RX_OFFSET_OFST; ++ ++ if (size) { ++ *len = CL_DESC_BUF_LEN(desc->ctrl) - PFE_PKT_HEADER_SZ - size; ++ *ofst = pfe_pkt_headroom + PFE_PKT_HEADER_SZ + size; ++ *priv_data = desc->data + PFE_PKT_HEADER_SZ; ++ } else { ++ *len = CL_DESC_BUF_LEN(desc->ctrl) - PFE_PKT_HEADER_SZ; ++ *ofst = pfe_pkt_headroom + PFE_PKT_HEADER_SZ; ++ *priv_data = NULL; ++ } ++ ++ } else { ++ *len = CL_DESC_BUF_LEN(desc->ctrl); ++ *ofst = pfe_pkt_headroom; ++ } ++ ++ desc->data = NULL; // Needed so we don't free a buffer/page twice on module_exit ++ smp_wmb(); ++ ++ desc->ctrl = CL_DESC_BUF_LEN(pfe_pkt_size) | CL_DESC_OWN; ++ inc_cl_idx(queue->read_idx); ++ } ++ ++ //spin_unlock_irqrestore(&client->rx_lock, flags); ++ return pkt; ++} ++ ++static inline void hif_hdr_write(struct hif_hdr *pkt_hdr, unsigned int client_id, unsigned int qno, u32 client_ctrl) ++{ ++ /* Optimize the write since the destinaton may be non-cacheable */ ++ if (!((unsigned long)pkt_hdr & 0x3)) { ++ ((u32 *)pkt_hdr)[0] = (client_ctrl << 16) | (qno << 8) | client_id; ++ } else { ++ ((u16 *)pkt_hdr)[0] = (qno << 8) | (client_id & 0xFF); ++ ((u16 *)pkt_hdr)[1] = (client_ctrl & 0xFFFF); ++ } ++} ++ ++/*This function puts the given packet in the specific client queue */ ++void __hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void *data, unsigned int len, u32 client_ctrl, unsigned int flags, void *client_data) ++{ ++ struct hif_client_tx_queue *queue = &client->tx_q[qno]; ++ struct tx_queue_desc *desc = queue->base + queue->write_idx; ++ ++ //printk(KERN_INFO "%s\n",__func__); ++ ++ /* First buffer */ ++ if (flags & HIF_FIRST_BUFFER) ++ { ++ data -= sizeof(struct hif_hdr); ++ len += sizeof(struct hif_hdr); ++ ++ hif_hdr_write(data, client->id, qno, client_ctrl); ++ } ++ ++ desc->data = client_data; ++ desc->ctrl = CL_DESC_OWN | CL_DESC_FLAGS(flags); ++ ++ __hif_xmit_pkt(&pfe->hif, client->id, qno, data, len, flags); ++ ++ inc_cl_idx(queue->write_idx); ++ queue->tx_pending++; ++ queue->jiffies_last_packet = jiffies; ++ ++} ++ ++/*This function puts the given packet in the specific client queue */ ++int hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void *data, unsigned int len, u32 client_ctrl, void *client_data) ++{ ++ struct hif_client_tx_queue *queue = &client->tx_q[qno]; ++ struct tx_queue_desc *desc = queue->base + queue->write_idx; ++ ++ //printk(KERN_INFO "%s\n",__func__); ++ ++ if (queue->tx_pending < queue->size) { ++ /*Construct pkt header */ ++ ++ data -= sizeof(struct hif_hdr); ++ len += sizeof(struct hif_hdr); ++ ++ hif_hdr_write(data, client->id, qno, client_ctrl); ++ ++ desc->data = client_data; ++ desc->ctrl = CL_DESC_OWN | CL_DESC_FLAGS(HIF_FIRST_BUFFER | HIF_LAST_BUFFER | HIF_DATA_VALID); ++ ++ if (hif_xmit_pkt(&pfe->hif, client->id, qno, data, len)) ++ return 1; ++ ++ inc_cl_idx(queue->write_idx); ++ queue->tx_pending++; ++ queue->jiffies_last_packet = jiffies; ++ ++ return 0; ++ } ++ ++ dbg_print_info("%s Tx client %d qno %d is full\n",__func__, client->id, qno); ++ return 1; ++} ++ ++void *hif_lib_tx_get_next_complete(struct hif_client_s *client, int qno, unsigned int *flags, int count) ++{ ++ struct hif_client_tx_queue *queue = &client->tx_q[qno]; ++ struct tx_queue_desc *desc = queue->base + queue->read_idx; ++ ++ dbg_print_info("%s: qno : %d rd_indx: %d pending:%d\n",__func__, qno, queue->read_idx, queue->tx_pending); ++ ++ if (!queue->tx_pending ) ++ return NULL; ++ ++ if (queue->nocpy_flag && !queue->done_tmu_tx_pkts) { ++ u32 tmu_tx_pkts = be32_to_cpu(pe_dmem_read(TMU0_ID + client->id, virt_to_tmu_dmem(&tx_trans[qno]), 4)); ++ ++ if (queue->prev_tmu_tx_pkts > tmu_tx_pkts) ++ queue->done_tmu_tx_pkts = UINT_MAX - queue->prev_tmu_tx_pkts + tmu_tx_pkts; ++ else ++ queue->done_tmu_tx_pkts = tmu_tx_pkts - queue->prev_tmu_tx_pkts; ++ ++ queue->prev_tmu_tx_pkts = tmu_tx_pkts; ++ ++ if (!queue->done_tmu_tx_pkts) { ++ return NULL; ++ } ++ } ++ ++ if (desc->ctrl & CL_DESC_OWN) { ++ hif_tx_done_process(&pfe->hif, count); ++ ++ //Check again, if packets done in tx queue. ++ if (desc->ctrl & CL_DESC_OWN) ++ return NULL; ++ } ++ ++ inc_cl_idx(queue->read_idx); ++ queue->tx_pending--; ++ ++ *flags = CL_DESC_GET_FLAGS(desc->ctrl); ++ ++ if (queue->done_tmu_tx_pkts && (*flags & HIF_LAST_BUFFER)) ++ queue->done_tmu_tx_pkts--; ++ ++ ++ return desc->data; ++} ++ ++//FIXME: TMU queues length mapping needs to be declared in shared PFE/PFE_CTRL header ++static void hif_lib_tmu_credit_init(struct pfe *pfe) ++{ ++ int i, q; ++ ++ for (i = 0; i < NUM_GEMAC_SUPPORT; i++) ++ for (q = 0; q < emac_txq_cnt; q++) { ++ pfe->tmu_credit.tx_credit_max[i][q] = (q == 0) ? DEFAULT_Q0_QDEPTH : DEFAULT_MAX_QDEPTH; ++ pfe->tmu_credit.tx_credit[i][q] = pfe->tmu_credit.tx_credit_max[i][q]; ++ } ++} ++/** __hif_lib_update_credit ++ * ++ * @param[in] client hif client context ++ * @param[in] queue queue number in match with TMU ++ */ ++void __hif_lib_update_credit(struct hif_client_s *client, unsigned int queue) ++{ ++ unsigned int tmu_tx_packets, tmp; ++ ++ if (tx_qos) { ++ tmu_tx_packets = be32_to_cpu(pe_dmem_read(TMU0_ID + client->id, virt_to_tmu_dmem(&tx_trans[queue]), 4)); ++ ++ // tx_packets counter overflowed ++ if (tmu_tx_packets > pfe->tmu_credit.tx_packets[client->id][queue]) { ++ tmp = UINT_MAX - tmu_tx_packets + pfe->tmu_credit.tx_packets[client->id][queue]; ++ pfe->tmu_credit.tx_credit[client->id][queue] = pfe->tmu_credit.tx_credit_max[client->id][queue] - tmp; ++ } ++ // TMU tx <= pfe_eth tx, normal case or both OF since last time ++ else ++ pfe->tmu_credit.tx_credit[client->id][queue] = pfe->tmu_credit.tx_credit_max[client->id][queue] - (pfe->tmu_credit.tx_packets[client->id][queue] - tmu_tx_packets); ++ } ++} ++ ++/** hif_lib_update_credit ++ * ++ * @param[in] client hif client context ++ * @param[in] queue queue number in match with TMU ++ */ ++void hif_lib_update_credit(struct hif_client_s *client, unsigned int queue) ++{ ++ spin_lock_bh(&pfe->hif.tx_lock); ++ __hif_lib_update_credit(client, queue); ++ spin_unlock_bh(&pfe->hif.tx_lock); ++} ++ ++int pfe_hif_lib_init(struct pfe *pfe) ++{ ++ int rc; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ if (lro_mode) { ++ page_mode = 1; ++ pfe_pkt_size = min(PAGE_SIZE, MAX_PFE_PKT_SIZE); ++ pfe_pkt_headroom = 0; ++ } else { ++ page_mode = 0; ++ pfe_pkt_size = PFE_PKT_SIZE; ++ pfe_pkt_headroom = PFE_PKT_HEADROOM; ++ } ++ ++ if (tx_qos) ++ emac_txq_cnt = EMAC_TXQ_CNT / 2; ++ else ++ emac_txq_cnt = EMAC_TXQ_CNT; ++ ++ hif_lib_tmu_credit_init(pfe); ++ pfe->hif.shm = &ghif_shm; ++ rc = pfe_hif_shm_init(pfe->hif.shm); ++ ++ return rc; ++} ++ ++ ++void pfe_hif_lib_exit(struct pfe *pfe) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ pfe_hif_shm_clean(pfe->hif.shm); ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hif_lib.h +@@ -0,0 +1,219 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_HIF_LIB_H_ ++#define _PFE_HIF_LIB_H_ ++ ++#include "pfe_hif.h" ++ ++#ifdef HIF_LIB_DEBUG ++#define dbg_print_info( fmt, args...) \ ++ printk(KERN_INFO fmt, ##args) ++#else ++#define dbg_print_info( fmt, args...) ++#endif ++ ++#define HIF_CL_REQ_TIMEOUT 10 ++ ++#if defined(CONFIG_COMCERTO_ZONE_DMA_NCNB) ++#define GFP_DMA_PFE (GFP_DMA_NCNB | __GFP_NOWARN) ++#else ++#define GFP_DMA_PFE 0 ++#endif ++ ++enum { ++ REQUEST_CL_REGISTER = 0, ++ REQUEST_CL_UNREGISTER, ++ HIF_REQUEST_MAX ++}; ++ ++enum { ++ EVENT_HIGH_RX_WM = 0, /* Event to indicate that client rx queue is reached water mark level */ ++ EVENT_RX_PKT_IND, /* Event to indicate that, packet recieved for client */ ++ EVENT_TXDONE_IND, /* Event to indicate that, packet tx done for client */ ++ HIF_EVENT_MAX ++}; ++ ++/*structure to store client queue info */ ++ ++/*structure to store client queue info */ ++struct hif_client_rx_queue { ++ struct rx_queue_desc *base; ++ u32 size; ++ u32 read_idx; ++ u32 write_idx; ++}; ++ ++struct hif_client_tx_queue { ++ struct tx_queue_desc *base; ++ u32 size; ++ u32 read_idx; ++ u32 write_idx; ++ u32 tx_pending; ++ unsigned long jiffies_last_packet; ++ u32 nocpy_flag; ++ u32 prev_tmu_tx_pkts; ++ u32 done_tmu_tx_pkts; ++ u32 cur_tso_hdr_p; ++ int tso_buf_cnt; ++}; ++ ++struct hif_client_s ++{ ++ int id; ++ int tx_qn; ++ int rx_qn; ++ void *rx_qbase; ++ void *tx_qbase; ++ /* FIXME tx/rx_qsize fields can be removed after per queue depth is supported*/ ++ int tx_qsize; ++ int rx_qsize; ++ int cpu_id; ++ ++// spinlock_t rx_lock; ++ struct hif_client_tx_queue tx_q[HIF_CLIENT_QUEUES_MAX]; ++ struct hif_client_rx_queue rx_q[HIF_CLIENT_QUEUES_MAX]; ++ int (*event_handler)(void *priv, int event, int data); ++ unsigned long queue_mask[HIF_EVENT_MAX]; ++ struct pfe *pfe; ++ void *priv; ++}; ++ ++ ++/* Client specific shared memory ++ * It contains number of Rx/Tx queues, base addresses and queue sizes */ ++struct hif_client_shm { ++ u32 ctrl; /*0-7: number of Rx queues, 8-15: number of tx queues */ ++ unsigned long rx_qbase; /*Rx queue base address */ ++ u32 rx_qsize; /*each Rx queue size, all Rx queues are of same size */ ++ unsigned long tx_qbase; /* Tx queue base address */ ++ u32 tx_qsize; /*each Tx queue size, all Tx queues are of same size */ ++}; ++ ++/*Client shared memory ctrl bit description */ ++#define CLIENT_CTRL_RX_Q_CNT_OFST 0 ++#define CLIENT_CTRL_TX_Q_CNT_OFST 8 ++#define CLIENT_CTRL_RX_Q_CNT(ctrl) (((ctrl) >> CLIENT_CTRL_RX_Q_CNT_OFST) & 0xFF) ++#define CLIENT_CTRL_TX_Q_CNT(ctrl) (((ctrl) >> CLIENT_CTRL_TX_Q_CNT_OFST) & 0xFF) ++ ++ ++ ++/*Shared memory used to communicate between HIF driver and host/client drivers ++ * Before starting the hif driver rx_buf_pool ans rx_buf_pool_cnt should be ++ * initialized with host buffers and buffers count in the pool. ++ * rx_buf_pool_cnt should be >= HIF_RX_DESC_NT. ++ * ++ */ ++struct hif_shm { ++ u32 rx_buf_pool_cnt; /*Number of rx buffers available*/ ++ void *rx_buf_pool[HIF_RX_DESC_NT];/*Rx buffers required to initialize HIF rx descriptors */ ++ unsigned long gClient_status[2]; /*Global client status bit mask */ ++ u32 hif_qfull; /*TODO Client-id that caused for the TMU3 queue stop */ ++ u32 hif_qresume; /*TODO */ ++ struct hif_client_shm client[HIF_CLIENTS_MAX]; /* Client specific shared memory */ ++}; ++ ++ ++#define CL_DESC_OWN (1 << 31) /* This sets owner ship to HIF driver */ ++#define CL_DESC_LAST (1 << 30) /* This indicates last packet for multi buffers handling */ ++#define CL_DESC_FIRST (1 << 29) /* This indicates first packet for multi buffers handling */ ++#define CL_DESC_BUF_LEN(x) ((x) & 0xFFFF) ++#define CL_DESC_FLAGS(x) (((x) & 0xF) << 16) ++#define CL_DESC_GET_FLAGS(x) (((x) >> 16) & 0xF) ++ ++struct rx_queue_desc { ++ void *data; ++ u32 ctrl; /*0-15bit len, 16-20bit flags, 31bit owner*/ ++ u32 client_ctrl; ++}; ++ ++struct tx_queue_desc { ++ void *data; ++ u32 ctrl; /*0-15bit len, 16-20bit flags, 31bit owner*/ ++}; ++ ++/* HIF Rx is not working properly for 2-byte aligned buffers and ++ * ip_header should be 4byte aligned for better iperformance. ++ * "ip_header = 64 + 6(hif_header) + 14 (MAC Header)" will be 4byte aligned. ++ */ ++#define PFE_PKT_HEADER_SZ sizeof(struct hif_hdr) ++#define PFE_BUF_SIZE 2048 /* must be big enough for headroom, pkt size and skb shared info */ ++#define PFE_PKT_HEADROOM 128 ++#define SKB_SHARED_INFO_SIZE 256 /* At least sizeof(struct skb_shared_info) bytes */ ++ ++//#define PFE_PKT_SIZE 1544 /* maximum ethernet packet size */ ++#define PFE_PKT_SIZE (PFE_BUF_SIZE - PFE_PKT_HEADROOM - SKB_SHARED_INFO_SIZE) /* maximum ethernet packet size after reassembly offload*/ ++#define MAX_L2_HDR_SIZE 14 /* Not correct for VLAN/PPPoE */ ++#define MAX_L3_HDR_SIZE 20 /* Not correct for IPv6 */ ++#define MAX_L4_HDR_SIZE 60 /* TCP with maximum options */ ++#define MAX_HDR_SIZE (MAX_L2_HDR_SIZE + MAX_L3_HDR_SIZE + MAX_L4_HDR_SIZE) ++#define MAX_WIFI_HDR_SIZE (MAX_L2_HDR_SIZE + MAX_L3_HDR_SIZE + 6) ++#define MAX_PFE_PKT_SIZE 16380UL /* Used in page mode to clamp packet size to the maximum supported by the hif hw interface (<16KiB) */ ++ ++extern unsigned int pfe_pkt_size; ++extern unsigned int pfe_pkt_headroom; ++extern unsigned int page_mode; ++extern unsigned int lro_mode; ++extern unsigned int tx_qos; ++extern unsigned int emac_txq_cnt; ++ ++int pfe_hif_lib_init(struct pfe *pfe); ++void pfe_hif_lib_exit(struct pfe *pfe); ++int hif_lib_client_register(struct hif_client_s *client); ++int hif_lib_client_unregister(struct hif_client_s *client); ++void __hif_lib_xmit_tso_hdr(struct hif_client_s *client, unsigned int qno, u32 client_ctrl, unsigned int ip_off, unsigned int ip_id, unsigned int ip_len, unsigned int tcp_off, unsigned int tcp_seq); ++void __hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void *data, unsigned int len, u32 client_ctrl, unsigned int flags, void *client_data); ++int hif_lib_xmit_pkt(struct hif_client_s *client, unsigned int qno, void *data, unsigned int len, u32 client_ctrl, void *client_data); ++void hif_lib_indicate_client(int cl_id, int event, int data); ++int hif_lib_event_handler_start( struct hif_client_s *client, int event, int data ); ++int hif_lib_tmu_queue_start( struct hif_client_s *client, int qno ); ++int hif_lib_tmu_queue_stop( struct hif_client_s *client, int qno ); ++void *hif_lib_tx_get_next_complete(struct hif_client_s *client, int qno, unsigned int *flags, int count); ++void *hif_lib_receive_pkt(struct hif_client_s *client, int qno, int *len, int *ofst, unsigned int *rx_ctrl, unsigned int *desc_ctrl, void **priv_data); ++void hif_lib_update_credit(struct hif_client_s *client, unsigned int qno); ++void __hif_lib_update_credit(struct hif_client_s *client, unsigned int queue); ++void hif_lib_set_rx_cpu_affinity(struct hif_client_s *client, int cpu_id); ++void hif_lib_set_tx_queue_nocpy(struct hif_client_s *client, int qno, int enable); ++static inline int hif_lib_tx_avail(struct hif_client_s *client, unsigned int qno) ++{ ++ struct hif_client_tx_queue *queue = &client->tx_q[qno]; ++ ++ return (queue->size - queue->tx_pending); ++} ++ ++static inline int hif_lib_get_tx_wrIndex(struct hif_client_s *client, unsigned int qno) ++{ ++ struct hif_client_tx_queue *queue = &client->tx_q[qno]; ++ ++ return queue->write_idx; ++} ++ ++ ++static inline int hif_lib_tx_pending(struct hif_client_s *client, unsigned int qno) ++{ ++ struct hif_client_tx_queue *queue = &client->tx_q[qno]; ++ ++ return queue->tx_pending; ++} ++ ++#define hif_lib_tx_credit_avail(pfe, id, qno) pfe->tmu_credit.tx_credit[id][qno] ++#define hif_lib_tx_credit_max(pfe, id, qno) pfe->tmu_credit.tx_credit_max[id][qno] ++#define hif_lib_tx_credit_use(pfe, id, qno, credit) do {if (tx_qos) {pfe->tmu_credit.tx_credit[id][qno]-= credit; pfe->tmu_credit.tx_packets[id][qno]+=credit;}} while (0) ++ ++#endif /* _PFE_HIF_LIB_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hw.c +@@ -0,0 +1,188 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#include "pfe_mod.h" ++#include "pfe_hw.h" ++ ++/* Functions to handle most of pfe hw register initialization */ ++ ++int pfe_hw_init(struct pfe *pfe, int resume) ++{ ++ CLASS_CFG class_cfg = { ++ .pe_sys_clk_ratio = PE_SYS_CLK_RATIO, ++ .route_table_baseaddr = pfe->ddr_phys_baseaddr + ROUTE_TABLE_BASEADDR, ++ .route_table_hash_bits = ROUTE_TABLE_HASH_BITS, ++ }; ++ ++ TMU_CFG tmu_cfg = { ++ .pe_sys_clk_ratio = PE_SYS_CLK_RATIO, ++ .llm_base_addr = pfe->ddr_phys_baseaddr + TMU_LLM_BASEADDR, ++ .llm_queue_len = TMU_LLM_QUEUE_LEN, ++ }; ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ UTIL_CFG util_cfg = { ++ .pe_sys_clk_ratio = PE_SYS_CLK_RATIO, ++ }; ++#endif ++ ++ BMU_CFG bmu1_cfg = { ++ .baseaddr = CBUS_VIRT_TO_PFE(LMEM_BASE_ADDR + BMU1_LMEM_BASEADDR), ++ .count = BMU1_BUF_COUNT, ++ .size = BMU1_BUF_SIZE, ++ }; ++ ++ BMU_CFG bmu2_cfg = { ++ .baseaddr = DDR_PHYS_TO_PFE(pfe->ddr_phys_baseaddr + BMU2_DDR_BASEADDR), ++ .count = BMU2_BUF_COUNT, ++ .size = BMU2_BUF_SIZE, ++ }; ++ ++ GPI_CFG egpi1_cfg = { ++ .lmem_rtry_cnt = EGPI1_LMEM_RTRY_CNT, ++ .tmlf_txthres = EGPI1_TMLF_TXTHRES, ++ .aseq_len = EGPI1_ASEQ_LEN, ++ }; ++ ++ GPI_CFG egpi2_cfg = { ++ .lmem_rtry_cnt = EGPI2_LMEM_RTRY_CNT, ++ .tmlf_txthres = EGPI2_TMLF_TXTHRES, ++ .aseq_len = EGPI2_ASEQ_LEN, ++ }; ++ ++#if defined(CONFIG_PLATFORM_C2000) ++ GPI_CFG egpi3_cfg = { ++ .lmem_rtry_cnt = EGPI3_LMEM_RTRY_CNT, ++ .tmlf_txthres = EGPI3_TMLF_TXTHRES, ++ .aseq_len = EGPI3_ASEQ_LEN, ++ }; ++#endif ++ ++ GPI_CFG hgpi_cfg = { ++ .lmem_rtry_cnt = HGPI_LMEM_RTRY_CNT, ++ .tmlf_txthres = HGPI_TMLF_TXTHRES, ++ .aseq_len = HGPI_ASEQ_LEN, ++ }; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++#if defined(CONFIG_PLATFORM_LS1012A) && !defined(LS1012A_PFE_RESET_WA) ++ /* LS1012A needs this to make PE work correctly */ ++ writel(0x3, CLASS_PE_SYS_CLK_RATIO); ++ writel(0x3, TMU_PE_SYS_CLK_RATIO); ++ writel(0x3, UTIL_PE_SYS_CLK_RATIO); ++ udelay(10); ++#endif ++ ++ printk(KERN_INFO "CLASS version: %x\n", readl(CLASS_VERSION)); ++ printk(KERN_INFO "TMU version: %x\n", readl(TMU_VERSION)); ++ ++ printk(KERN_INFO "BMU1 version: %x\n", readl(BMU1_BASE_ADDR + BMU_VERSION)); ++ printk(KERN_INFO "BMU2 version: %x\n", readl(BMU2_BASE_ADDR + BMU_VERSION)); ++#if defined(CONFIG_PLATFORM_C2000) ++ printk(KERN_INFO "EMAC1 network cfg: %x\n", readl(EMAC1_BASE_ADDR + EMAC_NETWORK_CONFIG)); ++ printk(KERN_INFO "EMAC2 network cfg: %x\n", readl(EMAC2_BASE_ADDR + EMAC_NETWORK_CONFIG)); ++#if !defined(CONFIG_PLATFORM_PCI) ++ printk(KERN_INFO "EMAC3 network cfg: %x\n", readl(EMAC3_BASE_ADDR + EMAC_NETWORK_CONFIG)); ++#endif ++#else ++ //TODO print MTIP config ++#endif ++ ++ printk(KERN_INFO "EGPI1 version: %x\n", readl(EGPI1_BASE_ADDR + GPI_VERSION)); ++ printk(KERN_INFO "EGPI2 version: %x\n", readl(EGPI2_BASE_ADDR + GPI_VERSION)); ++#if !defined(CONFIG_PLATFORM_PCI) && !defined(CONFIG_PLATFORM_LS1012A) ++ printk(KERN_INFO "EGPI3 version: %x\n", readl(EGPI3_BASE_ADDR + GPI_VERSION)); ++#endif ++ printk(KERN_INFO "HGPI version: %x\n", readl(HGPI_BASE_ADDR + GPI_VERSION)); ++ ++#if !defined(CONFIG_PLATFORM_PCI) ++ printk(KERN_INFO "GPT version: %x\n", readl(CBUS_GPT_VERSION)); ++#endif ++ ++ printk(KERN_INFO "HIF version: %x\n", readl(HIF_VERSION)); ++ printk(KERN_INFO "HIF NOPCY version: %x\n", readl(HIF_NOCPY_VERSION)); ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ printk(KERN_INFO "UTIL version: %x\n", readl(UTIL_VERSION)); ++#endif ++ while(!(readl(TMU_CTRL) & ECC_MEM_INIT_DONE)) ; ++ ++ hif_rx_disable(); ++ hif_tx_disable(); ++ ++ bmu_init(BMU1_BASE_ADDR, &bmu1_cfg); ++ ++ printk(KERN_INFO "bmu_init(1) done\n"); ++ ++ bmu_init(BMU2_BASE_ADDR, &bmu2_cfg); ++ ++ printk(KERN_INFO "bmu_init(2) done\n"); ++ ++ class_cfg.resume = resume ? 1 : 0; ++ ++ class_init(&class_cfg); ++ ++ printk(KERN_INFO "class_init() done\n"); ++ ++ tmu_init(&tmu_cfg); ++ ++ printk(KERN_INFO "tmu_init() done\n"); ++#if !defined(CONFIG_UTIL_DISABLED) ++ util_init(&util_cfg); ++ ++ printk(KERN_INFO "util_init() done\n"); ++#endif ++ gpi_init(EGPI1_BASE_ADDR, &egpi1_cfg); ++ ++ printk(KERN_INFO "gpi_init(1) done\n"); ++ ++ gpi_init(EGPI2_BASE_ADDR, &egpi2_cfg); ++ ++ printk(KERN_INFO "gpi_init(2) done\n"); ++#if !defined(CONFIG_PLATFORM_PCI) && !defined(CONFIG_PLATFORM_LS1012A) ++ gpi_init(EGPI3_BASE_ADDR, &egpi3_cfg); ++ ++ printk(KERN_INFO "gpi_init(3) done\n"); ++#endif ++ gpi_init(HGPI_BASE_ADDR, &hgpi_cfg); ++ ++ printk(KERN_INFO "gpi_init(hif) done\n"); ++ ++ bmu_enable(BMU1_BASE_ADDR); ++ ++ printk(KERN_INFO "bmu_enable(1) done\n"); ++ ++ bmu_enable(BMU2_BASE_ADDR); ++ ++ printk(KERN_INFO "bmu_enable(2) done\n"); ++ ++ return 0; ++} ++ ++void pfe_hw_exit(struct pfe *pfe) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ bmu_disable(BMU1_BASE_ADDR); ++ bmu_reset(BMU1_BASE_ADDR); ++ ++ bmu_disable(BMU2_BASE_ADDR); ++ bmu_reset(BMU2_BASE_ADDR); ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_hw.h +@@ -0,0 +1,32 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_HW_H_ ++#define _PFE_HW_H_ ++ ++#if !defined(CONFIG_PLATFORM_PCI) ++#define PE_SYS_CLK_RATIO 1 /* SYS/AXI = 250MHz, HFE = 500MHz */ ++#else ++#define PE_SYS_CLK_RATIO 0 /* SYS = 40MHz, HFE = 40MHz */ ++#endif ++ ++int pfe_hw_init(struct pfe *pfe, int resume); ++void pfe_hw_exit(struct pfe *pfe); ++ ++#endif /* _PFE_HW_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_ls1012a_platform.c +@@ -0,0 +1,341 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#include <linux/module.h> ++#include <linux/device.h> ++#include <linux/of_net.h> ++#include <linux/of_address.h> ++#include <linux/platform_device.h> ++#include <linux/slab.h> ++#include <linux/clk.h> ++#include <linux/mfd/syscon.h> ++#include <linux/regmap.h> ++ ++ ++#include "pfe_mod.h" ++ ++struct comcerto_pfe_platform_data pfe_platform_data; ++ ++ ++ ++static int pfe_get_gemac_if_proprties(struct device_node *parent, int port, int if_cnt, ++ struct comcerto_pfe_platform_data *pdata) ++{ ++ struct device_node *gem = NULL, *phy = NULL; ++ int size; ++ int ii = 0, phy_id = 0; ++ const u32 *addr; ++ const void *mac_addr; ++ ++ for (ii = 0; ii < if_cnt; ii++) { ++ gem = of_get_next_child(parent, gem); ++ if (!gem) ++ goto err; ++ addr = of_get_property(gem, "reg", &size); ++ if (addr && (be32_to_cpup(addr) == port)) ++ break; ++ } ++ ++ if (ii >= if_cnt) { ++ printk(KERN_ERR "%s:%d Failed to find interface = %d\n", __func__, __LINE__, if_cnt); ++ goto err; ++ } ++ ++ pdata->comcerto_eth_pdata[port].gem_id = port; ++ ++ mac_addr = of_get_mac_address(gem); ++ ++ if (mac_addr) { ++ memcpy(pdata->comcerto_eth_pdata[port].mac_addr, mac_addr, ETH_ALEN); ++ } ++ ++ if ((pdata->comcerto_eth_pdata[port].mii_config = of_get_phy_mode(gem)) < 0) ++ printk(KERN_ERR "%s:%d Incorrect Phy mode....\n", __func__, __LINE__); ++ ++ ++ addr = of_get_property(gem, "fsl,gemac-bus-id", &size); ++ if (!addr) ++ printk(KERN_ERR "%s:%d Invalid gemac-bus-id....\n", __func__, __LINE__); ++ else ++ pdata->comcerto_eth_pdata[port].bus_id = be32_to_cpup(addr); ++ ++ addr = of_get_property(gem, "fsl,gemac-phy-id", &size); ++ if (!addr) ++ printk(KERN_ERR "%s:%d Invalid gemac-phy-id....\n", __func__, __LINE__); ++ else ++ phy_id = pdata->comcerto_eth_pdata[port].phy_id = be32_to_cpup(addr); ++ ++ addr = of_get_property(gem, "fsl,mdio-mux-val", &size); ++ if (!addr) ++ printk(KERN_ERR "%s: Invalid mdio-mux-val....\n", __func__); ++ else ++ phy_id = pdata->comcerto_eth_pdata[port].mdio_muxval= be32_to_cpup(addr); ++ ++ ++ addr = of_get_property(gem, "fsl,pfe-phy-if-flags", &size); ++ if (!addr) ++ printk(KERN_ERR "%s:%d Invalid pfe-phy-if-flags....\n", __func__, __LINE__); ++ else ++ pdata->comcerto_eth_pdata[port].phy_flags = be32_to_cpup(addr); ++ ++ addr = of_get_property(gem, "fsl,pfe-gemac-mode", &size); ++ if (!addr) ++ printk(KERN_ERR "%s:%d Invalid pfe-gemac-mode....\n", __func__, __LINE__); ++ else ++ pdata->comcerto_eth_pdata[port].gemac_mode = be32_to_cpup(addr); ++ ++ ++ /* If PHY is enabled, read mdio properties */ ++ if (pdata->comcerto_eth_pdata[port].phy_flags & GEMAC_NO_PHY) ++ goto done; ++ ++ phy = of_get_next_child(gem, NULL); ++ ++ addr = of_get_property(phy, "reg", &size); ++ ++ if (!addr) ++ printk(KERN_ERR "%s:%d Invalid phy enable flag....\n", __func__, __LINE__); ++ else ++ pdata->comcerto_mdio_pdata[port].enabled = be32_to_cpup(addr); ++ ++ addr = of_get_property (phy, "fsl,mdio-phy-mask", &size); ++ if (!addr) ++ printk(KERN_ERR "%s:%d Unable to read mdio-phy-mask....\n", __func__, __LINE__); ++ else ++ pdata->comcerto_mdio_pdata[port].phy_mask= be32_to_cpup(addr); ++ pdata->comcerto_mdio_pdata[port].irq[0] = PHY_POLL; ++ ++done: ++ ++ return 0; ++ ++err: ++ return -1; ++} ++/** ++ * pfe_platform_probe - ++ * ++ * ++ */ ++static int pfe_platform_probe(struct platform_device *pdev) ++{ ++ struct resource res; ++ int ii, rc, interface_count = 0, size = 0; ++ const u32 *prop; ++ struct device_node *np; ++ ++ np = pdev->dev.of_node; ++ ++ if (!np) { ++ printk(KERN_ERR "Invalid device node\n"); ++ return -EINVAL; ++ } ++ ++ pfe = kzalloc(sizeof(struct pfe), GFP_KERNEL); ++ if (!pfe) { ++ rc = -ENOMEM; ++ goto err_alloc; ++ } ++ ++ platform_set_drvdata(pdev, pfe); ++ ++ dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32)); ++ ++ if (of_address_to_resource(np, 2, &res)) ++ { ++ rc = -ENOMEM; ++ printk(KERN_ERR "failed to get ddr resource\n"); ++ goto err_ddr; ++ } ++ ++ ++ pfe->ddr_phys_baseaddr = res.start; ++ pfe->ddr_size = resource_size(&res); ++ ++ //pfe->ddr_baseaddr = ioremap(res.start, resource_size(&res)); ++ pfe->ddr_baseaddr = phys_to_virt(res.start); ++ if (!pfe->ddr_baseaddr) { ++ printk(KERN_ERR "ioremap() ddr failed\n"); ++ rc = -ENOMEM; ++ goto err_ddr; ++ } ++ ++ /*printk("%s:%d : DDR Res : Phy addr:len = %x:%x Mapped addr : %x\n", __func__, __LINE__, ++ pfe->ddr_phys_baseaddr, pfe->ddr_size, pfe->ddr_baseaddr);*/ ++ ++ pfe->scfg = syscon_regmap_lookup_by_phandle(pdev->dev.of_node,"fsl,pfe-scfg"); ++ if (IS_ERR(pfe->scfg)) { ++ dev_err(&pdev->dev, "No syscfg phandle specified\n"); ++ return PTR_ERR(pfe->scfg); ++ } ++ /*printk("%s scfg %p\n",__func__,pfe->scfg);*/ ++ ++ ++#if 1 ++ if (!(pfe->cbus_baseaddr = of_iomap(np, 1))) ++ { ++ rc = -ENOMEM; ++ printk(KERN_ERR "failed to get axi resource\n"); ++ goto err_axi; ++ } ++ ++ /*printk("%s:%d : AXI Mapped addr : %lx\n", __func__, __LINE__, pfe->cbus_baseaddr); ++ printk("%s:%d : AXI Mapped addr : phys %lx\n", __func__, __LINE__, virt_to_phys(pfe->cbus_baseaddr));*/ ++#else ++ ++ if (of_address_to_resource(np, 1, &res)) ++ { ++ rc = -ENOMEM; ++ printk(KERN_ERR "failed to get AXI resource\n"); ++ goto err_iram; ++ } ++ pfe->cbus_baseaddr = ioremap(res.start, resource_size(&res)); ++ if (!pfe->cbus_baseaddr) { ++ printk(KERN_INFO "ioremap() AXI failed %lx %x\n", res.start, resource_size(&res)); ++ rc = -ENOMEM; ++ goto err_iram; ++ } ++ printk("%s:%d : AXI Mapped addr : %x PHY addr = %x\n", __func__, __LINE__, pfe->cbus_baseaddr, res.start); ++#endif ++ ++ pfe->hif_irq = platform_get_irq(pdev, 0); ++ if (pfe->hif_irq < 0) { ++ printk(KERN_ERR "platform_get_irq_byname(hif) failed\n"); ++ rc = pfe->hif_irq; ++ goto err_hif_irq; ++ } ++ /*printk("hif_irq: %d \n", pfe->hif_irq);*/ ++ ++ /* Read interface count */ ++ prop = of_get_property(np, "fsl,pfe-num-interfaces", &size); ++ if (!prop) { ++ printk(KERN_ERR "Failed to read number of interfaces\n"); ++ rc = -ENXIO; ++ goto err_prop; ++ } ++ ++ interface_count = be32_to_cpup(prop); ++ /*printk(KERN_INFO "%s:%d Number of interfaces : %d\n", __func__, __LINE__, interface_count);*/ ++ if (interface_count <= 0) { ++ printk(KERN_ERR "No ethernet interface count : %d\n", interface_count); ++ rc = -ENXIO; ++ goto err_prop; ++ } ++ ++ for (ii = 0; ii < interface_count; ii++) { ++ pfe_get_gemac_if_proprties(np, ii, interface_count, &pfe_platform_data); ++ } ++ ++ ++ pfe->dev = &pdev->dev; ++ ++ pfe->dev->platform_data = &pfe_platform_data; ++ ++ //FIXME get the correct clock from dts ++ pfe->ctrl.sys_clk = 250000; // save sys_clk value as KHz ++ ++ rc = pfe_probe(pfe); ++ if (rc < 0) ++ goto err_probe; ++ ++ return 0; ++ ++err_probe: ++err_prop: ++ /*TODO complet the code */ ++err_hif_irq: ++ iounmap(pfe->cbus_baseaddr); ++ ++err_axi: ++ iounmap(pfe->ddr_baseaddr); ++ ++err_ddr: ++ platform_set_drvdata(pdev, NULL); ++ ++ kfree(pfe); ++ ++err_alloc: ++ return rc; ++} ++ ++ ++/** ++ * pfe_platform_remove - ++ * ++ * ++ */ ++static int pfe_platform_remove(struct platform_device *pdev) ++{ ++ struct pfe *pfe = platform_get_drvdata(pdev); ++ int rc; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ rc = pfe_remove(pfe); ++ ++ iounmap(pfe->cbus_baseaddr); ++ iounmap(pfe->ddr_baseaddr); ++ ++ platform_set_drvdata(pdev, NULL); ++ ++ kfree(pfe); ++ ++ return rc; ++} ++ ++static struct of_device_id pfe_match[] = { ++ { ++ .compatible = "fsl,pfe", ++ }, ++ {}, ++}; ++MODULE_DEVICE_TABLE(of, pfe_match); ++ ++static struct platform_driver pfe_platform_driver = { ++ .probe = pfe_platform_probe, ++ .remove = pfe_platform_remove, ++ .driver = { ++ .name = "pfe", ++ .of_match_table = pfe_match, ++ }, ++}; ++ ++#if 0 ++static int __init pfe_module_init(void) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ return platform_driver_register(&pfe_platform_driver); ++} ++ ++ ++static void __exit pfe_module_exit(void) ++{ ++ platform_driver_unregister(&pfe_platform_driver); ++ ++ printk(KERN_INFO "%s\n", __func__); ++} ++module_init(pfe_module_init); ++module_exit(pfe_module_exit); ++#endif ++ ++module_platform_driver(pfe_platform_driver); ++MODULE_LICENSE("GPL"); ++MODULE_DESCRIPTION("PFE Ethernet driver"); ++MODULE_AUTHOR("NXP DNCPE"); +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_mod.c +@@ -0,0 +1,140 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#include <linux/dma-mapping.h> ++#include "pfe_mod.h" ++ ++struct pfe *pfe; ++ ++/** ++ * pfe_probe - ++ * ++ * ++ */ ++int pfe_probe(struct pfe *pfe) ++{ ++ int rc; ++ ++ ++ if (DDR_MAX_SIZE > pfe->ddr_size) { ++ printk(KERN_ERR "%s: required DDR memory (%x) above platform ddr memory (%x)\n", __func__, DDR_MAX_SIZE, pfe->ddr_size); ++ rc = -ENOMEM; ++ goto err_hw; ++ } ++ ++ if (((int) (pfe->ddr_phys_baseaddr + BMU2_DDR_BASEADDR) & (8*SZ_1M - 1)) != 0) { ++ printk(KERN_ERR "%s: BMU2 base address (0x%x) must be aligned on 8MB boundary\n", __func__, (int) pfe->ddr_phys_baseaddr + BMU2_DDR_BASEADDR); ++ rc = -ENOMEM; ++ goto err_hw; ++ } ++ ++ ++ printk(KERN_INFO "cbus_baseaddr: %lx, ddr_baseaddr: %lx, ddr_phys_baseaddr: %lx, ddr_size: %x\n", ++ (unsigned long)pfe->cbus_baseaddr, (unsigned long)pfe->ddr_baseaddr, ++ pfe->ddr_phys_baseaddr, pfe->ddr_size); ++ ++ pfe_lib_init(pfe->cbus_baseaddr, pfe->ddr_baseaddr, pfe->ddr_phys_baseaddr, pfe->ddr_size); ++ ++ rc = pfe_hw_init(pfe, 0); ++ if (rc < 0) ++ goto err_hw; ++ ++ rc = pfe_hif_lib_init(pfe); ++ if (rc < 0) ++ goto err_hif_lib; ++ ++ rc = pfe_hif_init(pfe); ++ if (rc < 0) ++ goto err_hif; ++ ++ rc = pfe_firmware_init(pfe); ++ if (rc < 0) ++ goto err_firmware; ++ ++ rc = pfe_ctrl_init(pfe); ++ if (rc < 0) ++ goto err_ctrl; ++ ++ rc = pfe_eth_init(pfe); ++ if (rc < 0) ++ goto err_eth; ++ ++ rc = pfe_sysfs_init(pfe); ++ if(rc < 0) ++ goto err_sysfs; ++ ++ rc = pfe_debugfs_init(pfe); ++ if (rc < 0) ++ goto err_debugfs; ++ ++ return 0; ++ ++err_debugfs: ++ pfe_sysfs_exit(pfe); ++ ++err_sysfs: ++ pfe_eth_exit(pfe); ++ ++err_eth: ++ pfe_ctrl_exit(pfe); ++ ++err_ctrl: ++ pfe_firmware_exit(pfe); ++ ++err_firmware: ++ pfe_hif_exit(pfe); ++ ++err_hif: ++ pfe_hif_lib_exit(pfe); ++ ++err_hif_lib: ++ pfe_hw_exit(pfe); ++ ++err_hw: ++ return rc; ++} ++ ++ ++/** ++ * pfe_remove - ++ * ++ * ++ */ ++int pfe_remove(struct pfe *pfe) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ pfe_debugfs_exit(pfe); ++ ++ pfe_sysfs_exit(pfe); ++ ++ pfe_eth_exit(pfe); ++ ++ pfe_ctrl_exit(pfe); ++ ++ pfe_firmware_exit(pfe); ++ ++ pfe_hif_exit(pfe); ++ ++ pfe_hif_lib_exit(pfe); ++ ++ pfe_hw_exit(pfe); ++ ++ return 0; ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_mod.h +@@ -0,0 +1,163 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_MOD_H_ ++#define _PFE_MOD_H_ ++ ++#include <linux/device.h> ++#include <linux/elf.h> ++ ++struct pfe; ++ ++#include "config.h" ++#include "pfe_hw.h" ++#include "pfe_firmware.h" ++#include "pfe_ctrl.h" ++#include "pfe_hif.h" ++#include "pfe_hif_lib.h" ++#include "pfe_eth.h" ++#include "pfe_sysfs.h" ++#include "pfe_perfmon.h" ++#include "pfe_debugfs.h" ++ ++struct pfe_tmu_credit { ++ /* Number of allowed TX packet in-flight, matches TMU queue size */ ++ unsigned int tx_credit[NUM_GEMAC_SUPPORT][EMAC_TXQ_CNT]; ++ unsigned int tx_credit_max[NUM_GEMAC_SUPPORT][EMAC_TXQ_CNT]; ++ unsigned int tx_packets[NUM_GEMAC_SUPPORT][EMAC_TXQ_CNT]; ++}; ++ ++struct pfe { ++ struct regmap *scfg; ++ unsigned long ddr_phys_baseaddr; ++ void *ddr_baseaddr; ++ unsigned int ddr_size; ++ void *cbus_baseaddr; ++ void *apb_baseaddr; ++ unsigned long iram_phys_baseaddr; ++ void *iram_baseaddr; ++ unsigned long ipsec_phys_baseaddr; ++ void *ipsec_baseaddr; ++ int hif_irq; ++ int hif_client_irq; ++ struct device *dev; ++ struct dentry *dentry; ++ struct pfe_ctrl ctrl; ++ struct pfe_hif hif; ++ struct pfe_eth eth; ++ struct hif_client_s *hif_client[HIF_CLIENTS_MAX]; ++#if defined(CFG_DIAGS) ++ struct pfe_diags diags; ++#endif ++ struct pfe_tmu_credit tmu_credit; ++ struct pfe_cpumon cpumon; ++ struct pfe_memmon memmon; ++ int wake; ++ struct clk * hfe_clock; ++}; ++ ++extern struct pfe *pfe; ++ ++int pfe_probe(struct pfe *pfe); ++int pfe_remove(struct pfe *pfe); ++ ++#ifndef SZ_1K ++#define SZ_1K 1024 ++#endif ++ ++#ifndef SZ_1M ++#define SZ_1M (1024 * 1024) ++#endif ++ ++/* DDR Mapping */ ++#if !defined(CONFIG_PLATFORM_PCI) ++#define ROUTE_TABLE_BASEADDR 0 ++#define ROUTE_TABLE_HASH_BITS 15 /**< 32K entries */ ++#define ROUTE_TABLE_SIZE ((1 << ROUTE_TABLE_HASH_BITS) * CLASS_ROUTE_SIZE) ++#define BMU2_DDR_BASEADDR (ROUTE_TABLE_BASEADDR + ROUTE_TABLE_SIZE) ++#define BMU2_BUF_COUNT (4096 - 256) /**< This is to get a total DDR size of 12MiB */ ++#define BMU2_DDR_SIZE (DDR_BUF_SIZE * BMU2_BUF_COUNT) ++#define UTIL_CODE_BASEADDR (BMU2_DDR_BASEADDR + BMU2_DDR_SIZE) ++#define UTIL_CODE_SIZE (128 * SZ_1K) ++#define UTIL_DDR_DATA_BASEADDR (UTIL_CODE_BASEADDR + UTIL_CODE_SIZE) ++#define UTIL_DDR_DATA_SIZE (64 * SZ_1K) ++#define CLASS_DDR_DATA_BASEADDR (UTIL_DDR_DATA_BASEADDR + UTIL_DDR_DATA_SIZE) ++#define CLASS_DDR_DATA_SIZE (32 * SZ_1K) ++#define TMU_DDR_DATA_BASEADDR (CLASS_DDR_DATA_BASEADDR + CLASS_DDR_DATA_SIZE) ++#define TMU_DDR_DATA_SIZE (32 * SZ_1K) ++#define TMU_LLM_BASEADDR (TMU_DDR_DATA_BASEADDR + TMU_DDR_DATA_SIZE) ++#define TMU_LLM_QUEUE_LEN (8 * 512) /**< Must be power of two and at least 16 * 8 = 128 bytes */ ++#define TMU_LLM_SIZE (4 * 16 * TMU_LLM_QUEUE_LEN) /**< (4 TMU's x 16 queues x queue_len) */ ++ ++#define DDR_MAX_SIZE (TMU_LLM_BASEADDR + TMU_LLM_SIZE) ++ ++#else ++ ++#define UTIL_CODE_BASEADDR 0 ++#if defined(CONFIG_UTIL_DISABLED) ++#define UTIL_CODE_SIZE (0 * SZ_1K) ++#else ++#define UTIL_CODE_SIZE (8 * SZ_1K) ++#endif ++#define UTIL_DDR_DATA_BASEADDR (UTIL_CODE_BASEADDR + UTIL_CODE_SIZE) ++#define UTIL_DDR_DATA_SIZE (0 * SZ_1K) ++#define CLASS_DDR_DATA_BASEADDR (UTIL_DDR_DATA_BASEADDR + UTIL_DDR_DATA_SIZE) ++#define CLASS_DDR_DATA_SIZE (0 * SZ_1K) ++#define TMU_DDR_DATA_BASEADDR (CLASS_DDR_DATA_BASEADDR + CLASS_DDR_DATA_SIZE) ++#define TMU_DDR_DATA_SIZE (0 * SZ_1K) ++#define ROUTE_TABLE_BASEADDR (TMU_DDR_DATA_BASEADDR + TMU_DDR_DATA_SIZE) ++#define ROUTE_TABLE_HASH_BITS 5 /**< 32 entries */ ++#define ROUTE_TABLE_SIZE ((1 << ROUTE_TABLE_HASH_BITS) * CLASS_ROUTE_SIZE) ++#define BMU2_DDR_BASEADDR (ROUTE_TABLE_BASEADDR + ROUTE_TABLE_SIZE) ++#define BMU2_BUF_COUNT 16 ++#define BMU2_DDR_SIZE (DDR_BUF_SIZE * BMU2_BUF_COUNT) ++#define TMU_LLM_BASEADDR (BMU2_DDR_BASEADDR + BMU2_DDR_SIZE) ++#define TMU_LLM_QUEUE_LEN (16 * 8) /**< Must be power of two and at least 16 * 8 = 128 bytes */ ++#define TMU_LLM_SIZE (4 * 16 * TMU_LLM_QUEUE_LEN) /**< (4 TMU's x 16 queues x queue_len) */ ++#define HIF_DESC_BASEADDR (TMU_LLM_BASEADDR + TMU_LLM_SIZE) ++#define HIF_RX_DESC_SIZE (16*HIF_RX_DESC_NT) ++#define HIF_TX_DESC_SIZE (16*HIF_TX_DESC_NT) ++#define HIF_DESC_SIZE (HIF_RX_DESC_SIZE + HIF_TX_DESC_SIZE) ++#define HIF_RX_PKT_DDR_BASEADDR (HIF_DESC_BASEADDR + HIF_DESC_SIZE) ++#define HIF_RX_PKT_DDR_SIZE (HIF_RX_DESC_NT * DDR_BUF_SIZE) ++#define HIF_TX_PKT_DDR_BASEADDR (HIF_RX_PKT_DDR_BASEADDR + HIF_RX_PKT_DDR_SIZE) ++#define HIF_TX_PKT_DDR_SIZE (HIF_TX_DESC_NT * DDR_BUF_SIZE) ++#define ROUTE_BASEADDR (HIF_TX_PKT_DDR_BASEADDR + HIF_TX_PKT_DDR_SIZE) ++#define ROUTE_SIZE (2 * CLASS_ROUTE_SIZE) ++ ++#define DDR_MAX_SIZE (ROUTE_BASEADDR + ROUTE_SIZE) ++ ++#define PFE_HOST_TO_PCI(addr) (((u32)addr)- ((u32)DDR_BASE_ADDR)) ++#define PFE_PCI_TO_HOST(addr) (((u32)addr)+ ((u32)DDR_BASE_ADDR)) ++#endif ++ ++/* IRAM Mapping */ ++#define IPSEC_IRAM_BASEADDR 0 ++#define IPSEC_IRAM_SIZE 0x2000 ++ ++/* LMEM Mapping */ ++#define BMU1_LMEM_BASEADDR 0 ++#define BMU1_BUF_COUNT 256 ++#define BMU1_LMEM_SIZE (LMEM_BUF_SIZE * BMU1_BUF_COUNT) ++#define IPSEC_LMEM_BASEADDR (BMU1_LMEM_BASEADDR + BMU1_LMEM_SIZE) ++#define IPSEC_LMEM_SIZE (30 * 1024) ++ ++ ++ ++#endif /* _PFE_MOD_H */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_perfmon.c +@@ -0,0 +1,175 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++/* PFE performance monitoring functions */ ++ ++#include "pfe_ctrl_hal.h" ++#include "pfe_perfmon.h" ++ ++static TIMER_ENTRY cpumon_timer; ++ ++u32 CLASS_DMEM_SH2(cpu_ticks[2]); ++u32 TMU_DMEM_SH2(cpu_ticks[2]); ++#if !defined(CONFIG_UTIL_DISABLED) ++u32 UTIL_DMEM_SH2(cpu_ticks[2]); ++#endif ++ ++#define compute_active_pct(total_ticks, active_ticks) ((active_ticks * 100 + (total_ticks >> 1)) / total_ticks) ++ ++static void cpumon_timer_handler(void) ++{ ++ int id; ++ u32 dmem_addr; ++ u32 ticks[2]; ++ u32 total, active; ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ struct pfe_cpumon *cpumon = &pfe->cpumon; ++ ++ // Process class PE's ++ total = active = 0; ++ dmem_addr = virt_to_class_dmem(&class_cpu_ticks[0]); ++ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) ++ { ++ cpumon->cpu_usage_pct[id] = 0; ++ if (pe_sync_stop(ctrl, (1 << id)) < 0) ++ continue; ++ ticks[0] = be32_to_cpu(pe_dmem_read(id, dmem_addr, 4)); ++ ticks[1] = be32_to_cpu(pe_dmem_read(id, dmem_addr + 4, 4)); ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ pe_dmem_write(id, 0, dmem_addr + 4, 4); ++ pe_start(ctrl, (1 << id)); ++ ticks[0] >>= 8; // divide both values by 256, so multiply by 100 won't overflow ++ ticks[1] >>= 8; ++ total += ticks[0]; ++ active += ticks[1]; ++ if (ticks[0] != 0) ++ cpumon->cpu_usage_pct[id] = compute_active_pct(ticks[0], ticks[1]); ++ } ++ if (total != 0) ++ cpumon->class_usage_pct = compute_active_pct(total, active); ++ else ++ cpumon->class_usage_pct = 0; ++ ++ // Process TMU PE's ++ total = active = 0; ++ dmem_addr = virt_to_tmu_dmem(&tmu_cpu_ticks[0]); ++ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) ++ { ++#if defined(CONFIG_PLATFORM_LS1012A) ++ if(id == TMU2_ID) continue; ++#endif ++ cpumon->cpu_usage_pct[id] = 0; ++ if (pe_sync_stop(ctrl, (1 << id)) < 0) ++ continue; ++ ticks[0] = be32_to_cpu(pe_dmem_read(id, dmem_addr, 4)); ++ ticks[1] = be32_to_cpu(pe_dmem_read(id, dmem_addr + 4, 4)); ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ pe_dmem_write(id, 0, dmem_addr + 4, 4); ++ pe_start(ctrl, (1 << id)); ++ ticks[0] >>= 8; // divide both values by 256, so multiply by 100 won't overflow ++ ticks[1] >>= 8; ++ if (ticks[0] != 0) ++ cpumon->cpu_usage_pct[id] = compute_active_pct(ticks[0], ticks[1]); ++ } ++#if !defined(CONFIG_UTIL_DISABLED) ++ // Process Util PE ++ dmem_addr = virt_to_util_dmem(&util_cpu_ticks[0]); ++ cpumon->cpu_usage_pct[UTIL_ID] = 0; ++ if (pe_sync_stop(ctrl, (1 << UTIL_ID)) < 0) ++ return; ++ ticks[0] = be32_to_cpu(pe_dmem_read(UTIL_ID, dmem_addr, 4)); ++ ticks[1] = be32_to_cpu(pe_dmem_read(UTIL_ID, dmem_addr + 4, 4)); ++ pe_dmem_write(UTIL_ID, 0, dmem_addr, 4); ++ pe_dmem_write(UTIL_ID, 0, dmem_addr + 4, 4); ++ pe_start(ctrl, (1 << UTIL_ID)); ++ ticks[0] >>= 8; // divide both values by 256, so multiply by 100 won't overflow ++ ticks[1] >>= 8; ++ if (ticks[0] != 0) ++ cpumon->cpu_usage_pct[UTIL_ID] = compute_active_pct(ticks[0], ticks[1]); ++#endif ++} ++ ++static int pfe_cpumon_init(struct pfe *pfe) ++{ ++ timer_init(&cpumon_timer, cpumon_timer_handler); ++ timer_add(&cpumon_timer, CT_CPUMON_INTERVAL); ++ return 0; ++} ++ ++static void pfe_cpumon_exit(struct pfe *pfe) ++{ ++ timer_del(&cpumon_timer); ++} ++ ++ ++/*********************************************************************************/ ++ ++// Memory monitor functions ++ ++void * pfe_kmalloc(size_t size, int flags) ++{ ++ struct pfe_memmon *memmon = &pfe->memmon; ++ void *ptr; ++ ptr = kmalloc(size, flags); ++ if (ptr) ++ memmon->kernel_memory_allocated += ksize(ptr); ++ return ptr; ++} ++ ++void * pfe_kzalloc(size_t size, int flags) ++{ ++ struct pfe_memmon *memmon = &pfe->memmon; ++ void *ptr; ++ ptr = kzalloc(size, flags); ++ if (ptr) ++ memmon->kernel_memory_allocated += ksize(ptr); ++ return ptr; ++} ++ ++void pfe_kfree(void *ptr) ++{ ++ struct pfe_memmon *memmon = &pfe->memmon; ++ memmon->kernel_memory_allocated -= ksize(ptr); ++ kfree(ptr); ++} ++ ++static int pfe_memmon_init(struct pfe *pfe) ++{ ++ return 0; ++} ++ ++static void pfe_memmon_exit(struct pfe *pfe) ++{ ++} ++ ++/*********************************************************************************/ ++ ++ ++int pfe_perfmon_init(struct pfe *pfe) ++{ ++ pfe_cpumon_init(pfe); ++ pfe_memmon_init(pfe); ++ return 0; ++} ++ ++void pfe_perfmon_exit(struct pfe *pfe) ++{ ++ pfe_cpumon_exit(pfe); ++ pfe_memmon_exit(pfe); ++} +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_perfmon.h +@@ -0,0 +1,41 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_PERFMON_H_ ++#define _PFE_PERFMON_H_ ++ ++#define CT_CPUMON_INTERVAL (1 * TIMER_TICKS_PER_SEC) ++ ++struct pfe_cpumon { ++ u32 cpu_usage_pct[MAX_PE]; ++ u32 class_usage_pct; ++}; ++ ++struct pfe_memmon { ++ u32 kernel_memory_allocated; ++}; ++ ++void * pfe_kmalloc(size_t size, int flags); ++void * pfe_kzalloc(size_t size, int flags); ++void pfe_kfree(void *ptr); ++ ++int pfe_perfmon_init(struct pfe *pfe); ++void pfe_perfmon_exit(struct pfe *pfe); ++ ++#endif /* _PFE_PERFMON_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_platform.c +@@ -0,0 +1,358 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#include <linux/module.h> ++#include <linux/platform_device.h> ++#include <linux/slab.h> ++#include <linux/clk.h> ++ ++#include "pfe_mod.h" ++ ++/** ++ * pfe_platform_probe - ++ * ++ * ++ */ ++static int pfe_platform_probe(struct platform_device *pdev) ++{ ++ struct resource *r; ++ int rc; ++ struct clk *clk_axi; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ pfe = kzalloc(sizeof(struct pfe), GFP_KERNEL); ++ if (!pfe) { ++ rc = -ENOMEM; ++ goto err_alloc; ++ } ++ ++ platform_set_drvdata(pdev, pfe); ++ ++ r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ddr"); ++ if (!r) { ++ printk(KERN_INFO "platform_get_resource_byname(ddr) failed\n"); ++ rc = -ENXIO; ++ goto err_ddr; ++ } ++ ++ pfe->ddr_phys_baseaddr = r->start; ++ pfe->ddr_size = resource_size(r); ++ ++ pfe->ddr_baseaddr = ioremap(r->start, resource_size(r)); ++ if (!pfe->ddr_baseaddr) { ++ printk(KERN_INFO "ioremap() ddr failed\n"); ++ rc = -ENOMEM; ++ goto err_ddr; ++ } ++ ++ r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "axi"); ++ if (!r) { ++ printk(KERN_INFO "platform_get_resource_byname(axi) failed\n"); ++ rc = -ENXIO; ++ goto err_axi; ++ } ++ ++ pfe->cbus_baseaddr = ioremap(r->start, resource_size(r)); ++ if (!pfe->cbus_baseaddr) { ++ printk(KERN_INFO "ioremap() axi failed\n"); ++ rc = -ENOMEM; ++ goto err_axi; ++ } ++ ++ r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "apb"); ++ if (!r) { ++ printk(KERN_INFO "platform_get_resource_byname(apb) failed\n"); ++ rc = -ENXIO; ++ goto err_apb; ++ } ++ ++ pfe->apb_baseaddr = ioremap(r->start, resource_size(r)); ++ if (!pfe->apb_baseaddr) { ++ printk(KERN_INFO "ioremap() apb failed\n"); ++ rc = -ENOMEM; ++ goto err_apb; ++ } ++ ++ r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "iram"); ++ if (!r) { ++ printk(KERN_INFO "platform_get_resource_byname(iram) failed\n"); ++ rc = -ENXIO; ++ goto err_iram; ++ } ++ ++ pfe->iram_phys_baseaddr = r->start; ++ pfe->iram_baseaddr = ioremap(r->start, resource_size(r)); ++ if (!pfe->iram_baseaddr) { ++ printk(KERN_INFO "ioremap() iram failed\n"); ++ rc = -ENOMEM; ++ goto err_iram; ++ } ++ ++ r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ipsec"); ++ if (!r) { ++ printk(KERN_INFO "platform_get_resource_byname(ipsec) failed\n"); ++ rc = -ENXIO; ++ goto err_ipsec; ++ } ++ ++ pfe->ipsec_phys_baseaddr = r->start; ++ /* Just map only initial 1MB , as its enough to access espah engine ++ */ ++ //pfe->ipsec_baseaddr = ioremap(r->start, resource_size(r)); ++ pfe->ipsec_baseaddr = ioremap(r->start, 1*1024*1024); ++ if (!pfe->ipsec_baseaddr) { ++ printk(KERN_INFO "ioremap() ipsec failed\n"); ++ rc = -ENOMEM; ++ goto err_ipsec; ++ } ++ ++ printk(KERN_INFO "ipsec: baseaddr :%x --- %x\n", (u32)pfe->ipsec_phys_baseaddr, (u32)pfe->ipsec_baseaddr); ++ ++ pfe->hif_irq = platform_get_irq_byname(pdev, "hif"); ++ if (pfe->hif_irq < 0) { ++ printk(KERN_INFO "platform_get_irq_byname(hif) failed\n"); ++ rc = pfe->hif_irq; ++ goto err_hif_irq; ++ } ++ ++#if 0 ++ pfe->hif_client_irq = platform_get_irq_byname(pdev, "hif_client"); ++ if (pfe->hif_client_irq < 0) { ++ printk(KERN_INFO "platform_get_irq_byname(hif_client) failed\n"); ++ rc = pfe->hif_client_irq; ++ goto err_hif_irq; ++ } ++#endif ++ ++ pfe->dev = &pdev->dev; ++ ++ ++ /* Get the system clock */ ++ clk_axi = clk_get(NULL,"axi"); ++ if (IS_ERR(clk_axi)) { ++ printk(KERN_INFO "clk_get call failed\n"); ++ rc = -ENXIO; ++ goto err_clk; ++ } ++ ++ /* HFE core clock */ ++ pfe->hfe_clock = clk_get(NULL, "hfe_core"); ++ if (IS_ERR(pfe->hfe_clock)) { ++ printk(KERN_INFO "clk_get call failed\n"); ++ rc = -ENXIO; ++ goto err_hfe_clock; ++ } ++ ++ clk_disable(pfe->hfe_clock); ++ c2000_block_reset(COMPONENT_PFE_SYS, 1); ++ mdelay(1); ++ c2000_block_reset(COMPONENT_PFE_SYS, 0); ++ clk_enable(pfe->hfe_clock); ++ ++ pfe->ctrl.clk_axi = clk_axi; ++ pfe->ctrl.sys_clk = clk_get_rate(clk_axi) / 1000; // save sys_clk value as KHz ++ ++ rc = pfe_probe(pfe); ++ if (rc < 0) ++ goto err_probe; ++ ++ return 0; ++ ++err_probe: ++ clk_put(pfe->hfe_clock); ++err_hfe_clock: ++ clk_put(clk_axi); ++err_clk: ++err_hif_irq: ++ iounmap(pfe->ipsec_baseaddr); ++err_ipsec: ++ iounmap(pfe->iram_baseaddr); ++err_iram: ++ iounmap(pfe->apb_baseaddr); ++ ++err_apb: ++ iounmap(pfe->cbus_baseaddr); ++ ++err_axi: ++ iounmap(pfe->ddr_baseaddr); ++ ++err_ddr: ++ platform_set_drvdata(pdev, NULL); ++ ++ kfree(pfe); ++ ++err_alloc: ++ return rc; ++} ++ ++ ++/** ++ * pfe_platform_remove - ++ * ++ * ++ */ ++static int pfe_platform_remove(struct platform_device *pdev) ++{ ++ struct pfe *pfe = platform_get_drvdata(pdev); ++ int rc; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ rc = pfe_remove(pfe); ++ ++ c2000_block_reset(COMPONENT_PFE_SYS, 1); ++ clk_disable(pfe->hfe_clock); ++ clk_put(pfe->hfe_clock); ++ clk_put(pfe->ctrl.clk_axi); ++ iounmap(pfe->ipsec_baseaddr); ++ iounmap(pfe->iram_baseaddr); ++ iounmap(pfe->apb_baseaddr); ++ iounmap(pfe->cbus_baseaddr); ++ iounmap(pfe->ddr_baseaddr); ++ ++ platform_set_drvdata(pdev, NULL); ++ ++ kfree(pfe); ++ ++ return rc; ++} ++ ++#ifdef CONFIG_PM ++ ++#ifdef CONFIG_PM_SLEEP ++static int pfe_platform_suspend(struct device *dev) ++{ ++ struct pfe *pfe = platform_get_drvdata(to_platform_device(dev)); ++ struct net_device *netdev; ++ int i; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ pfe->wake = 0; ++ ++ for (i = 0; i < (NUM_GEMAC_SUPPORT - 1); i++ ) { ++ netdev = pfe->eth.eth_priv[i]->dev; ++ ++ netif_device_detach(netdev); ++ ++ if (netif_running(netdev)) ++ if(pfe_eth_suspend(netdev)) ++ pfe->wake =1; ++ } ++ ++ /* Shutdown PFE only if we're not waking up the system */ ++ if (!pfe->wake) { ++ pfe_ctrl_suspend(&pfe->ctrl); ++ pfe_hif_exit(pfe); ++ pfe_hif_lib_exit(pfe); ++ ++ class_disable(); ++ tmu_disable(0xf); ++#if !defined(CONFIG_UTIL_DISABLED) ++ util_disable(); ++#endif ++ pfe_hw_exit(pfe); ++ c2000_block_reset(COMPONENT_PFE_SYS, 1); ++ clk_disable(pfe->hfe_clock); ++ } ++ ++ return 0; ++} ++ ++static int pfe_platform_resume(struct device *dev) ++{ ++ struct pfe *pfe = platform_get_drvdata(to_platform_device(dev)); ++ struct net_device *netdev; ++ int i; ++ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ if (!pfe->wake) { ++ /* Sequence follows VLSI recommendation (bug 71927) */ ++ c2000_block_reset(COMPONENT_PFE_SYS, 1); ++ mdelay(1); ++ c2000_block_reset(COMPONENT_PFE_SYS, 0); ++ clk_enable(pfe->hfe_clock); ++ ++ pfe_hw_init(pfe, 1); ++ pfe_hif_lib_init(pfe); ++ pfe_hif_init(pfe); ++#if !defined(CONFIG_UTIL_DISABLED) ++ util_enable(); ++#endif ++ tmu_enable(0xf); ++ class_enable(); ++ pfe_ctrl_resume(&pfe->ctrl); ++ } ++ ++ for(i = 0; i < (NUM_GEMAC_SUPPORT - 1); i++) { ++ netdev = pfe->eth.eth_priv[i]->dev; ++ ++ if (pfe->eth.eth_priv[i]->mii_bus) ++ pfe_eth_mdio_reset(pfe->eth.eth_priv[i]->mii_bus); ++ ++ if (netif_running(netdev)) ++ pfe_eth_resume(netdev); ++ ++ netif_device_attach(netdev); ++ } ++ return 0; ++} ++#else ++#define pfe_platform_suspend NULL ++#define pfe_platform_resume NULL ++#endif ++ ++static const struct dev_pm_ops pfe_platform_pm_ops = { ++ SET_SYSTEM_SLEEP_PM_OPS(pfe_platform_suspend, pfe_platform_resume) ++}; ++ ++#endif ++ ++static struct platform_driver pfe_platform_driver = { ++ .probe = pfe_platform_probe, ++ .remove = pfe_platform_remove, ++ .driver = { ++ .name = "pfe", ++#ifdef CONFIG_PM ++ .pm = &pfe_platform_pm_ops, ++#endif ++ }, ++}; ++ ++ ++static int __init pfe_module_init(void) ++{ ++ printk(KERN_INFO "%s\n", __func__); ++ ++ return platform_driver_register(&pfe_platform_driver); ++} ++ ++ ++static void __exit pfe_module_exit(void) ++{ ++ platform_driver_unregister(&pfe_platform_driver); ++ ++ printk(KERN_INFO "%s\n", __func__); ++} ++ ++MODULE_LICENSE("GPL"); ++module_init(pfe_module_init); ++module_exit(pfe_module_exit); +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_sysfs.c +@@ -0,0 +1,855 @@ ++/* ++ * (C) Copyright 2011 ++ * Author : Freescale Semiconductor, Inc. ++ * ++ * See file CREDITS for list of people who contributed to this ++ * project. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public License as ++ * published by the Free Software Foundation; either version 2 of ++ * the License, or (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, ++ * MA 02111-1307 USA ++ * */ ++ ++#include <linux/module.h> ++#include <linux/platform_device.h> ++ ++#include "pfe_mod.h" ++#include "pfe_ctrl_hal.h" ++ ++#define PE_EXCEPTION_DUMP_ADDRESS 0x1fa8 ++#define NUM_QUEUES 16 ++ ++static char register_name[20][5] = { ++ "EPC", "ECAS", "EID", "ED", ++ "r0", "r1", "r2", "r3", ++ "r4", "r5", "r6", "r7", ++ "r8", "r9", "r10", "r11", ++ "r12", "r13", "r14", "r15", ++}; ++ ++static char exception_name[14][20] = { ++ "Reset", ++ "HardwareFailure", ++ "NMI", ++ "InstBreakpoint", ++ "DataBreakpoint", ++ "Unsupported", ++ "PrivilegeViolation", ++ "InstBusError", ++ "DataBusError", ++ "AlignmentError", ++ "ArithmeticError", ++ "SystemCall", ++ "MemoryManagement", ++ "Interrupt", ++}; ++ ++static unsigned long class_do_clear = 0; ++static unsigned long tmu_do_clear = 0; ++#if !defined(CONFIG_UTIL_DISABLED) ++static unsigned long util_do_clear = 0; ++#endif ++ ++u32 qm_read_drop_stat(u32 tmu, u32 queue, u32 *total_drops, int do_reset); ++ ++static ssize_t display_pe_status(char *buf, int id, u32 dmem_addr, unsigned long do_clear) ++{ ++ ssize_t len = 0; ++ u32 val; ++ char statebuf[5]; ++ struct pfe_cpumon *cpumon = &pfe->cpumon; ++ u32 debug_indicator; ++ u32 debug[20]; ++ ++ *(u32 *)statebuf = pe_dmem_read(id, dmem_addr, 4); ++ dmem_addr += 4; ++ ++ statebuf[4] = '\0'; ++ len += sprintf(buf + len, "state=%4s ", statebuf); ++ ++ val = pe_dmem_read(id, dmem_addr, 4); ++ dmem_addr += 4; ++ len += sprintf(buf + len, "ctr=%08x ", cpu_to_be32(val)); ++ ++ val = pe_dmem_read(id, dmem_addr, 4); ++ if (do_clear && val) ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ dmem_addr += 4; ++ len += sprintf(buf + len, "rx=%u ", cpu_to_be32(val)); ++ ++ val = pe_dmem_read(id, dmem_addr, 4); ++ if (do_clear && val) ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ dmem_addr += 4; ++ if (id >= TMU0_ID && id <= TMU_MAX_ID) ++ len += sprintf(buf + len, "qstatus=%x", cpu_to_be32(val)); ++ else ++ len += sprintf(buf + len, "tx=%u", cpu_to_be32(val)); ++ ++ val = pe_dmem_read(id, dmem_addr, 4); ++ if (do_clear && val) ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ dmem_addr += 4; ++ if (val) ++ len += sprintf(buf + len, " drop=%u", cpu_to_be32(val)); ++ ++ len += sprintf(buf + len, " load=%d%%", cpumon->cpu_usage_pct[id]); ++ ++ len += sprintf(buf + len, "\n"); ++ ++ debug_indicator = pe_dmem_read(id, dmem_addr, 4); ++ dmem_addr += 4; ++ if (!strncmp((char *)&debug_indicator, "DBUG", 4)) ++ { ++ int j, last = 0; ++ for (j = 0; j < 16; j++) ++ { ++ debug[j] = pe_dmem_read(id, dmem_addr, 4); ++ if (debug[j]) ++ { ++ if (do_clear) ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ last = j + 1; ++ } ++ dmem_addr += 4; ++ } ++ for (j = 0; j < last; j++) ++ { ++ len += sprintf(buf + len, "%08x%s", cpu_to_be32(debug[j]), ++ (j & 0x7) == 0x7 || j == last - 1 ? "\n" : " "); ++ } ++ } ++ ++ if (!strncmp(statebuf, "DEAD", 4)) ++ { ++ u32 i, dump = PE_EXCEPTION_DUMP_ADDRESS; ++ ++ len += sprintf(buf + len, "Exception details:\n"); ++ for (i = 0; i < 20; i++) { ++ debug[i] = pe_dmem_read(id, dump, 4); ++ dump +=4; ++ if (i == 2) ++ len += sprintf(buf + len, "%4s = %08x (=%s) ", register_name[i], cpu_to_be32(debug[i]), exception_name[min((u32) cpu_to_be32(debug[i]), (u32)13)]); ++ else ++ len += sprintf(buf + len, "%4s = %08x%s", register_name[i], cpu_to_be32(debug[i]), ++ (i & 0x3) == 0x3 || i == 19 ? "\n" : " "); ++ } ++ } ++ ++ return len; ++} ++ ++static ssize_t class_phy_stats(char *buf, int phy) ++{ ++ ssize_t len = 0; ++ int off1 = phy * 0x28; ++ int off2 = phy * 0x10; ++ ++ if (phy == 3) ++ off1 = CLASS_PHY4_RX_PKTS - CLASS_PHY1_RX_PKTS; ++ ++ len += sprintf(buf + len, "phy: %d\n", phy); ++ len += sprintf(buf + len, " rx: %10u, tx: %10u, intf: %10u, ipv4: %10u, ipv6: %10u\n", ++ readl(CLASS_PHY1_RX_PKTS + off1), readl(CLASS_PHY1_TX_PKTS + off1), ++ readl(CLASS_PHY1_INTF_MATCH_PKTS + off1), readl(CLASS_PHY1_V4_PKTS + off1), ++ readl(CLASS_PHY1_V6_PKTS + off1)); ++ ++ len += sprintf(buf + len, " icmp: %10u, igmp: %10u, tcp: %10u, udp: %10u\n", ++ readl(CLASS_PHY1_ICMP_PKTS + off2), readl(CLASS_PHY1_IGMP_PKTS + off2), ++ readl(CLASS_PHY1_TCP_PKTS + off2), readl(CLASS_PHY1_UDP_PKTS + off2)); ++ ++ len += sprintf(buf + len, " err\n"); ++ len += sprintf(buf + len, " lp: %10u, intf: %10u, l3: %10u, chcksum: %10u, ttl: %10u\n", ++ readl(CLASS_PHY1_LP_FAIL_PKTS + off1), readl(CLASS_PHY1_INTF_FAIL_PKTS + off1), ++ readl(CLASS_PHY1_L3_FAIL_PKTS + off1), readl(CLASS_PHY1_CHKSUM_ERR_PKTS + off1), ++ readl(CLASS_PHY1_TTL_ERR_PKTS + off1)); ++ ++ return len; ++} ++ ++/** qm_read_drop_stat ++ * This function is used to read the drop statistics from the TMU ++ * hw drop counter. Since the hw counter is always cleared afer ++ * reading, this function maintains the previous drop count, and ++ * adds the new value to it. That value can be retrieved by ++ * passing a pointer to it with the total_drops arg. ++ * ++ * @param tmu TMU number (0 - 3) ++ * @param queue queue number (0 - 15) ++ * @param total_drops pointer to location to store total drops (or NULL) ++ * @param do_reset if TRUE, clear total drops after updating ++ * ++ */ ++ ++u32 qm_read_drop_stat(u32 tmu, u32 queue, u32 *total_drops, int do_reset) ++{ ++ static u32 qtotal[TMU_MAX_ID + 1][NUM_QUEUES]; ++ u32 val; ++ writel((tmu << 8) | queue, TMU_TEQ_CTRL); ++ writel((tmu << 8) | queue, TMU_LLM_CTRL); ++ val = readl(TMU_TEQ_DROP_STAT); ++ qtotal[tmu][queue] += val; ++ if (total_drops) ++ *total_drops = qtotal[tmu][queue]; ++ if (do_reset) ++ qtotal[tmu][queue] = 0; ++ return val; ++} ++ ++static ssize_t tmu_queue_stats(char *buf, int tmu, int queue) ++{ ++ ssize_t len = 0; ++ u32 drops; ++ ++ len += sprintf(buf + len, "%d-%02d, ", tmu, queue); ++ ++ drops = qm_read_drop_stat(tmu, queue, NULL, 0); ++ ++ /* Select queue */ ++ writel((tmu << 8) | queue, TMU_TEQ_CTRL); ++ writel((tmu << 8) | queue, TMU_LLM_CTRL); ++ ++ len += sprintf(buf + len, "(teq) drop: %10u, tx: %10u (llm) head: %08x, tail: %08x, drop: %10u\n", ++ drops, readl(TMU_TEQ_TRANS_STAT), ++ readl(TMU_LLM_QUE_HEADPTR), readl(TMU_LLM_QUE_TAILPTR), ++ readl(TMU_LLM_QUE_DROPCNT)); ++ ++ return len; ++} ++ ++ ++static ssize_t tmu_queues(char *buf, int tmu) ++{ ++ ssize_t len = 0; ++ int queue; ++ ++ for (queue = 0; queue < 16; queue++) ++ len += tmu_queue_stats(buf + len, tmu, queue); ++ ++ return len; ++} ++ ++static ssize_t tmu_ctx(char *buf, int tmu) ++{ ++ ssize_t len = 0; ++ int i; ++ u32 val, tmu_context_addr = TMU_CONTEXT_ADDR; ++ ++ len += sprintf(buf+len, " TMU %d \n", TMU0_ID+tmu); ++ for (i = 1; i <= 160 ; i++, tmu_context_addr += 4) ++ { ++ val = pe_dmem_read(TMU0_ID+tmu, tmu_context_addr , 4); ++ if (i == 5) ++ len += sprintf(buf+len, "\nShapers: Each shaper structure is 8 bytes and there are 10 shapers\n"); ++ ++ if (i == 25) ++ len += sprintf(buf+len, "\nScheduler: Each scheduler structure is 48 bytes and there are 8 schedulers\n"); ++ if (i == 121) ++ len += sprintf(buf+len, "\nQueue: Each queue structure is 2 bytes and there are 16 queues\n"); ++ ++ if (i == 129) ++ len += sprintf(buf+len, "\nqlenmasks array for 16 queues\n"); ++ if (i == 145) ++ len += sprintf(buf+len, "\nqresultmap array for 16 queues\n"); ++ if (i%8 == 0) ++ len += sprintf(buf+len, "%08x \n", cpu_to_be32(val)); ++ else ++ len += sprintf(buf+len, "%08x ", cpu_to_be32(val)); ++ } ++ ++ len += sprintf(buf+len, "\n"); ++ ++ return len; ++} ++ ++static ssize_t block_version(char *buf, void *addr) ++{ ++ ssize_t len = 0; ++ u32 val; ++ ++ val = readl(addr); ++ len += sprintf(buf + len, "revision: %x, version: %x, id: %x\n", (val >> 24) & 0xff, (val >> 16) & 0xff, val & 0xffff); ++ ++ return len; ++} ++ ++static ssize_t bmu(char *buf, int id, void *base) ++{ ++ ssize_t len = 0; ++ ++ len += sprintf(buf + len, "bmu: %d\n ", id); ++ ++ len += block_version(buf + len, base + BMU_VERSION); ++ ++ len += sprintf(buf + len, " buf size: %x\n", (1 << readl(base + BMU_BUF_SIZE))); ++ len += sprintf(buf + len, " buf count: %x\n", readl(base + BMU_BUF_CNT)); ++ len += sprintf(buf + len, " buf rem: %x\n", readl(base + BMU_REM_BUF_CNT)); ++ len += sprintf(buf + len, " buf curr: %x\n", readl(base + BMU_CURR_BUF_CNT)); ++ len += sprintf(buf + len, " free err: %x\n", readl(base + BMU_FREE_ERR_ADDR)); ++ ++ return len; ++} ++ ++static ssize_t gpi(char *buf, int id, void *base) ++{ ++ ssize_t len = 0; ++ u32 val; ++ ++ len += sprintf(buf + len, "gpi%d:\n ", id); ++ len += block_version(buf + len, base + GPI_VERSION); ++ ++ len += sprintf(buf + len, " tx under stick: %x\n", readl(base + GPI_FIFO_STATUS)); ++ val = readl(base + GPI_FIFO_DEBUG); ++ len += sprintf(buf + len, " tx pkts: %x\n", (val >> 23) & 0x3f); ++ len += sprintf(buf + len, " rx pkts: %x\n", (val >> 18) & 0x3f); ++ len += sprintf(buf + len, " tx bytes: %x\n", (val >> 9) & 0x1ff); ++ len += sprintf(buf + len, " rx bytes: %x\n", (val >> 0) & 0x1ff); ++ len += sprintf(buf + len, " overrun: %x\n", readl(base + GPI_OVERRUN_DROPCNT)); ++ ++ return len; ++} ++ ++static ssize_t pfe_set_class(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) ++{ ++ class_do_clear = simple_strtoul(buf, NULL, 0); ++ return count; ++} ++ ++static ssize_t pfe_show_class(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ int id; ++ u32 val; ++ struct pfe_cpumon *cpumon = &pfe->cpumon; ++ ++ len += block_version(buf + len, CLASS_VERSION); ++ ++ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) ++ { ++ len += sprintf(buf + len, "%d: ", id - CLASS0_ID); ++ ++ val = readl(CLASS_PE0_DEBUG + id * 4); ++ len += sprintf(buf + len, "pc=1%04x ", val & 0xffff); ++ ++ len += display_pe_status(buf + len, id, PESTATUS_ADDR_CLASS, class_do_clear); ++ } ++ len += sprintf(buf + len, "aggregate load=%d%%\n\n", cpumon->class_usage_pct); ++ ++ len += sprintf(buf + len, "pe status: 0x%x\n", readl(CLASS_PE_STATUS)); ++ len += sprintf(buf + len, "max buf cnt: 0x%x afull thres: 0x%x\n", readl(CLASS_MAX_BUF_CNT), readl(CLASS_AFULL_THRES)); ++ len += sprintf(buf + len, "tsq max cnt: 0x%x tsq fifo thres: 0x%x\n", readl(CLASS_TSQ_MAX_CNT), readl(CLASS_TSQ_FIFO_THRES)); ++ len += sprintf(buf + len, "state: 0x%x\n", readl(CLASS_STATE)); ++ ++ len += class_phy_stats(buf + len, 0); ++ len += class_phy_stats(buf + len, 1); ++ len += class_phy_stats(buf + len, 2); ++ len += class_phy_stats(buf + len, 3); ++ ++ return len; ++} ++ ++static ssize_t pfe_set_tmu(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) ++{ ++ tmu_do_clear = simple_strtoul(buf, NULL, 0); ++ return count; ++} ++ ++static ssize_t pfe_show_tmu(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ int id; ++ u32 val; ++ ++ len += block_version(buf + len, TMU_VERSION); ++ ++ for (id = TMU0_ID; id <= TMU_MAX_ID; id++) ++ { ++#if defined(CONFIG_PLATFORM_LS1012A) ++ if(id == TMU2_ID) continue; ++#endif ++ len += sprintf(buf + len, "%d: ", id - TMU0_ID); ++ ++ len += display_pe_status(buf + len, id, PESTATUS_ADDR_TMU, tmu_do_clear); ++ } ++ ++ len += sprintf(buf + len, "pe status: %x\n", readl(TMU_PE_STATUS)); ++ len += sprintf(buf + len, "inq fifo cnt: %x\n", readl(TMU_PHY_INQ_FIFO_CNT)); ++ val = readl(TMU_INQ_STAT); ++ len += sprintf(buf + len, "inq wr ptr: %x\n", val & 0x3ff); ++ len += sprintf(buf + len, "inq rd ptr: %x\n", val >> 10); ++ ++ ++ return len; ++} ++ ++ ++static unsigned long drops_do_clear = 0; ++static u32 CLASS_DMEM_SH2(drop_counter)[CLASS_NUM_DROP_COUNTERS]; ++#if !defined(CONFIG_UTIL_DISABLED) ++static u32 UTIL_DMEM_SH2(drop_counter)[UTIL_NUM_DROP_COUNTERS]; ++#endif ++ ++char *class_drop_description[CLASS_NUM_DROP_COUNTERS] = { ++ "ICC", ++ "Host Pkt Error", ++ "Rx Error", ++ "IPsec Outbound", ++ "IPsec Inbound", ++ "EXPT IPsec Error", ++ "Reassembly", ++ "Fragmenter", ++ "NAT-T", ++ "Socket", ++ "Multicast", ++ "NAT-PT", ++ "Tx Disabled", ++}; ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++char *util_drop_description[UTIL_NUM_DROP_COUNTERS] = { ++ "IPsec Outbound", ++ "IPsec Inbound", ++ "IPsec Rate Limiter", ++ "Fragmenter", ++ "Socket", ++ "Tx Disabled", ++ "Rx Error", ++}; ++#endif ++ ++static ssize_t pfe_set_drops(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) ++{ ++ drops_do_clear = simple_strtoul(buf, NULL, 0); ++ return count; ++} ++ ++static u32 tmu_drops[4][16]; ++static ssize_t pfe_show_drops(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ int id, dropnum; ++ int tmu, queue; ++ u32 val; ++ u32 dmem_addr; ++ int num_class_drops = 0, num_tmu_drops = 0, num_util_drops = 0; ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ memset(class_drop_counter, 0, sizeof(class_drop_counter)); ++ for (id = CLASS0_ID; id <= CLASS_MAX_ID; id++) ++ { ++ if (drops_do_clear) ++ pe_sync_stop(ctrl, (1 << id)); ++ for (dropnum = 0; dropnum < CLASS_NUM_DROP_COUNTERS; dropnum++) ++ { ++ dmem_addr = virt_to_class_dmem(&class_drop_counter[dropnum]); ++ val = be32_to_cpu(pe_dmem_read(id, dmem_addr, 4)); ++ class_drop_counter[dropnum] += val; ++ num_class_drops += val; ++ if (drops_do_clear) ++ pe_dmem_write(id, 0, dmem_addr, 4); ++ } ++ if (drops_do_clear) ++ pe_start(ctrl, (1 << id)); ++ } ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ if (drops_do_clear) ++ pe_sync_stop(ctrl, (1 << UTIL_ID)); ++ for (dropnum = 0; dropnum < UTIL_NUM_DROP_COUNTERS; dropnum++) ++ { ++ dmem_addr = virt_to_util_dmem(&util_drop_counter[dropnum]); ++ val = be32_to_cpu(pe_dmem_read(UTIL_ID, dmem_addr, 4)); ++ util_drop_counter[dropnum] = val; ++ num_util_drops += val; ++ if (drops_do_clear) ++ pe_dmem_write(UTIL_ID, 0, dmem_addr, 4); ++ } ++ if (drops_do_clear) ++ pe_start(ctrl, (1 << UTIL_ID)); ++#endif ++ for (tmu = 0; tmu < 4; tmu++) ++ { ++ for (queue = 0; queue < 16; queue++) ++ { ++ qm_read_drop_stat(tmu, queue, &tmu_drops[tmu][queue], drops_do_clear); ++ num_tmu_drops += tmu_drops[tmu][queue]; ++ } ++ } ++ ++ if (num_class_drops == 0 && num_util_drops == 0 && num_tmu_drops == 0) ++ len += sprintf(buf + len, "No PE drops\n\n"); ++ ++ if (num_class_drops > 0) ++ { ++ len += sprintf(buf + len, "Class PE drops --\n"); ++ for (dropnum = 0; dropnum < CLASS_NUM_DROP_COUNTERS; dropnum++) ++ { ++ if (class_drop_counter[dropnum] > 0) ++ len += sprintf(buf + len, " %s: %d\n", class_drop_description[dropnum], class_drop_counter[dropnum]); ++ } ++ len += sprintf(buf + len, "\n"); ++ } ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ if (num_util_drops > 0) ++ { ++ len += sprintf(buf + len, "Util PE drops --\n"); ++ for (dropnum = 0; dropnum < UTIL_NUM_DROP_COUNTERS; dropnum++) ++ { ++ if (util_drop_counter[dropnum] > 0) ++ len += sprintf(buf + len, " %s: %d\n", util_drop_description[dropnum], util_drop_counter[dropnum]); ++ } ++ len += sprintf(buf + len, "\n"); ++ } ++#endif ++ if (num_tmu_drops > 0) ++ { ++ len += sprintf(buf + len, "TMU drops --\n"); ++ for (tmu = 0; tmu < 4; tmu++) ++ { ++ for (queue = 0; queue < 16; queue++) ++ { ++ if (tmu_drops[tmu][queue] > 0) ++ len += sprintf(buf + len, " TMU%d-Q%d: %d\n", tmu, queue, tmu_drops[tmu][queue]); ++ } ++ } ++ len += sprintf(buf + len, "\n"); ++ } ++ ++ return len; ++} ++ ++static ssize_t pfe_show_tmu0_queues(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_queues(buf, 0); ++} ++ ++static ssize_t pfe_show_tmu1_queues(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_queues(buf, 1); ++} ++ ++static ssize_t pfe_show_tmu2_queues(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_queues(buf, 2); ++} ++ ++static ssize_t pfe_show_tmu3_queues(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_queues(buf, 3); ++} ++ ++static ssize_t pfe_show_tmu0_ctx(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_ctx(buf, 0); ++} ++static ssize_t pfe_show_tmu1_ctx(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_ctx(buf, 1); ++} ++static ssize_t pfe_show_tmu2_ctx(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_ctx(buf, 2); ++} ++ ++static ssize_t pfe_show_tmu3_ctx(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ return tmu_ctx(buf, 3); ++} ++ ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++static ssize_t pfe_set_util(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) ++{ ++ util_do_clear = simple_strtoul(buf, NULL, 0); ++ return count; ++} ++ ++static ssize_t pfe_show_util(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ struct pfe_ctrl *ctrl = &pfe->ctrl; ++ ++ ++ len += block_version(buf + len, UTIL_VERSION); ++ ++ pe_sync_stop(ctrl, (1 << UTIL_ID)); ++ len += display_pe_status(buf + len, UTIL_ID, PESTATUS_ADDR_UTIL, util_do_clear); ++ pe_start(ctrl, (1 << UTIL_ID)); ++ ++ len += sprintf(buf + len, "pe status: %x\n", readl(UTIL_PE_STATUS)); ++ len += sprintf(buf + len, "max buf cnt: %x\n", readl(UTIL_MAX_BUF_CNT)); ++ len += sprintf(buf + len, "tsq max cnt: %x\n", readl(UTIL_TSQ_MAX_CNT)); ++ ++ return len; ++} ++#endif ++ ++static ssize_t pfe_show_bmu(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ ++ len += bmu(buf + len, 1, BMU1_BASE_ADDR); ++ len += bmu(buf + len, 2, BMU2_BASE_ADDR); ++ ++ return len; ++} ++ ++static ssize_t pfe_show_hif(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ ++ len += sprintf(buf + len, "hif:\n "); ++ len += block_version(buf + len, HIF_VERSION); ++ ++ len += sprintf(buf + len, " tx curr bd: %x\n", readl(HIF_TX_CURR_BD_ADDR)); ++ len += sprintf(buf + len, " tx status: %x\n", readl(HIF_TX_STATUS)); ++ len += sprintf(buf + len, " tx dma status: %x\n", readl(HIF_TX_DMA_STATUS)); ++ ++ len += sprintf(buf + len, " rx curr bd: %x\n", readl(HIF_RX_CURR_BD_ADDR)); ++ len += sprintf(buf + len, " rx status: %x\n", readl(HIF_RX_STATUS)); ++ len += sprintf(buf + len, " rx dma status: %x\n", readl(HIF_RX_DMA_STATUS)); ++ ++ len += sprintf(buf + len, "hif nocopy:\n "); ++ len += block_version(buf + len, HIF_NOCPY_VERSION); ++ ++ len += sprintf(buf + len, " tx curr bd: %x\n", readl(HIF_NOCPY_TX_CURR_BD_ADDR)); ++ len += sprintf(buf + len, " tx status: %x\n", readl(HIF_NOCPY_TX_STATUS)); ++ len += sprintf(buf + len, " tx dma status: %x\n", readl(HIF_NOCPY_TX_DMA_STATUS)); ++ ++ len += sprintf(buf + len, " rx curr bd: %x\n", readl(HIF_NOCPY_RX_CURR_BD_ADDR)); ++ len += sprintf(buf + len, " rx status: %x\n", readl(HIF_NOCPY_RX_STATUS)); ++ len += sprintf(buf + len, " rx dma status: %x\n", readl(HIF_NOCPY_RX_DMA_STATUS)); ++ ++ return len; ++} ++ ++ ++static ssize_t pfe_show_gpi(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ ++ len += gpi(buf + len, 0, EGPI1_BASE_ADDR); ++ len += gpi(buf + len, 1, EGPI2_BASE_ADDR); ++#if !defined(CONFIG_PLATFORM_LS1012A) ++ len += gpi(buf + len, 2, EGPI3_BASE_ADDR); ++#endif ++ len += gpi(buf + len, 3, HGPI_BASE_ADDR); ++ ++ return len; ++} ++ ++static ssize_t pfe_show_pfemem(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ ssize_t len = 0; ++ struct pfe_memmon *memmon = &pfe->memmon; ++ ++ len += sprintf(buf + len, "Kernel Memory: %d Bytes (%d KB)\n", memmon->kernel_memory_allocated, (memmon->kernel_memory_allocated + 1023) / 1024); ++ ++ return len; ++} ++ ++#ifdef HIF_NAPI_STATS ++static ssize_t pfe_show_hif_napi_stats(struct device *dev, struct device_attribute *attr, char *buf) ++{ ++ struct platform_device *pdev = to_platform_device(dev); ++ struct pfe *pfe = platform_get_drvdata(pdev); ++ ssize_t len = 0; ++ ++ len += sprintf(buf + len, "sched: %u\n", pfe->hif.napi_counters[NAPI_SCHED_COUNT]); ++ len += sprintf(buf + len, "poll: %u\n", pfe->hif.napi_counters[NAPI_POLL_COUNT]); ++ len += sprintf(buf + len, "packet: %u\n", pfe->hif.napi_counters[NAPI_PACKET_COUNT]); ++ len += sprintf(buf + len, "budget: %u\n", pfe->hif.napi_counters[NAPI_FULL_BUDGET_COUNT]); ++ len += sprintf(buf + len, "desc: %u\n", pfe->hif.napi_counters[NAPI_DESC_COUNT]); ++ len += sprintf(buf + len, "full: %u\n", pfe->hif.napi_counters[NAPI_CLIENT_FULL_COUNT]); ++ ++ return len; ++} ++ ++static ssize_t pfe_set_hif_napi_stats(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) ++{ ++ struct platform_device *pdev = to_platform_device(dev); ++ struct pfe *pfe = platform_get_drvdata(pdev); ++ ++ memset(pfe->hif.napi_counters, 0, sizeof(pfe->hif.napi_counters)); ++ ++ return count; ++} ++ ++static DEVICE_ATTR(hif_napi_stats, 0644, pfe_show_hif_napi_stats, pfe_set_hif_napi_stats); ++#endif ++ ++ ++static DEVICE_ATTR(class, 0644, pfe_show_class, pfe_set_class); ++static DEVICE_ATTR(tmu, 0644, pfe_show_tmu, pfe_set_tmu); ++#if !defined(CONFIG_UTIL_DISABLED) ++static DEVICE_ATTR(util, 0644, pfe_show_util, pfe_set_util); ++#endif ++static DEVICE_ATTR(bmu, 0444, pfe_show_bmu, NULL); ++static DEVICE_ATTR(hif, 0444, pfe_show_hif, NULL); ++static DEVICE_ATTR(gpi, 0444, pfe_show_gpi, NULL); ++static DEVICE_ATTR(drops, 0644, pfe_show_drops, pfe_set_drops); ++static DEVICE_ATTR(tmu0_queues, 0444, pfe_show_tmu0_queues, NULL); ++static DEVICE_ATTR(tmu1_queues, 0444, pfe_show_tmu1_queues, NULL); ++static DEVICE_ATTR(tmu2_queues, 0444, pfe_show_tmu2_queues, NULL); ++static DEVICE_ATTR(tmu3_queues, 0444, pfe_show_tmu3_queues, NULL); ++static DEVICE_ATTR(tmu0_ctx, 0444, pfe_show_tmu0_ctx, NULL); ++static DEVICE_ATTR(tmu1_ctx, 0444, pfe_show_tmu1_ctx, NULL); ++static DEVICE_ATTR(tmu2_ctx, 0444, pfe_show_tmu2_ctx, NULL); ++static DEVICE_ATTR(tmu3_ctx, 0444, pfe_show_tmu3_ctx, NULL); ++static DEVICE_ATTR(pfemem, 0444, pfe_show_pfemem, NULL); ++ ++ ++int pfe_sysfs_init(struct pfe *pfe) ++{ ++ if (device_create_file(pfe->dev, &dev_attr_class)) ++ goto err_class; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu)) ++ goto err_tmu; ++ ++#if !defined(CONFIG_UTIL_DISABLED) ++ if (device_create_file(pfe->dev, &dev_attr_util)) ++ goto err_util; ++#endif ++ ++ if (device_create_file(pfe->dev, &dev_attr_bmu)) ++ goto err_bmu; ++ ++ if (device_create_file(pfe->dev, &dev_attr_hif)) ++ goto err_hif; ++ ++ if (device_create_file(pfe->dev, &dev_attr_gpi)) ++ goto err_gpi; ++ ++ if (device_create_file(pfe->dev, &dev_attr_drops)) ++ goto err_drops; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu0_queues)) ++ goto err_tmu0_queues; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu1_queues)) ++ goto err_tmu1_queues; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu2_queues)) ++ goto err_tmu2_queues; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu3_queues)) ++ goto err_tmu3_queues; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu0_ctx)) ++ goto err_tmu0_ctx; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu1_ctx)) ++ goto err_tmu1_ctx; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu2_ctx)) ++ goto err_tmu2_ctx; ++ ++ if (device_create_file(pfe->dev, &dev_attr_tmu3_ctx)) ++ goto err_tmu3_ctx; ++ ++ if (device_create_file(pfe->dev, &dev_attr_pfemem)) ++ goto err_pfemem; ++ ++#ifdef HIF_NAPI_STATS ++ if (device_create_file(pfe->dev, &dev_attr_hif_napi_stats)) ++ goto err_hif_napi_stats; ++#endif ++ ++ return 0; ++ ++#ifdef HIF_NAPI_STATS ++err_hif_napi_stats: ++ device_remove_file(pfe->dev, &dev_attr_pfemem); ++#endif ++ ++err_pfemem: ++ device_remove_file(pfe->dev, &dev_attr_tmu3_ctx); ++ ++err_tmu3_ctx: ++ device_remove_file(pfe->dev, &dev_attr_tmu2_ctx); ++ ++err_tmu2_ctx: ++ device_remove_file(pfe->dev, &dev_attr_tmu1_ctx); ++ ++err_tmu1_ctx: ++ device_remove_file(pfe->dev, &dev_attr_tmu0_ctx); ++ ++err_tmu0_ctx: ++ device_remove_file(pfe->dev, &dev_attr_tmu3_queues); ++ ++err_tmu3_queues: ++ device_remove_file(pfe->dev, &dev_attr_tmu2_queues); ++ ++err_tmu2_queues: ++ device_remove_file(pfe->dev, &dev_attr_tmu1_queues); ++ ++err_tmu1_queues: ++ device_remove_file(pfe->dev, &dev_attr_tmu0_queues); ++ ++err_tmu0_queues: ++ device_remove_file(pfe->dev, &dev_attr_drops); ++ ++err_drops: ++ device_remove_file(pfe->dev, &dev_attr_gpi); ++ ++err_gpi: ++ device_remove_file(pfe->dev, &dev_attr_hif); ++ ++err_hif: ++ device_remove_file(pfe->dev, &dev_attr_bmu); ++ ++err_bmu: ++#if !defined(CONFIG_UTIL_DISABLED) ++ device_remove_file(pfe->dev, &dev_attr_util); ++ ++err_util: ++#endif ++ device_remove_file(pfe->dev, &dev_attr_tmu); ++ ++err_tmu: ++ device_remove_file(pfe->dev, &dev_attr_class); ++ ++err_class: ++ return -1; ++} ++ ++ ++void pfe_sysfs_exit(struct pfe *pfe) ++{ ++#ifdef HIF_NAPI_STATS ++ device_remove_file(pfe->dev, &dev_attr_hif_napi_stats); ++#endif ++ ++ device_remove_file(pfe->dev, &dev_attr_pfemem); ++ device_remove_file(pfe->dev, &dev_attr_tmu3_ctx); ++ device_remove_file(pfe->dev, &dev_attr_tmu2_ctx); ++ device_remove_file(pfe->dev, &dev_attr_tmu1_ctx); ++ device_remove_file(pfe->dev, &dev_attr_tmu0_ctx); ++ device_remove_file(pfe->dev, &dev_attr_tmu3_queues); ++ device_remove_file(pfe->dev, &dev_attr_tmu2_queues); ++ device_remove_file(pfe->dev, &dev_attr_tmu1_queues); ++ device_remove_file(pfe->dev, &dev_attr_tmu0_queues); ++ device_remove_file(pfe->dev, &dev_attr_drops); ++ device_remove_file(pfe->dev, &dev_attr_gpi); ++ device_remove_file(pfe->dev, &dev_attr_hif); ++ device_remove_file(pfe->dev, &dev_attr_bmu); ++#if !defined(CONFIG_UTIL_DISABLED) ++ device_remove_file(pfe->dev, &dev_attr_util); ++#endif ++ device_remove_file(pfe->dev, &dev_attr_tmu); ++ device_remove_file(pfe->dev, &dev_attr_class); ++} ++ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/pfe_sysfs.h +@@ -0,0 +1,34 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PFE_SYSFS_H_ ++#define _PFE_SYSFS_H_ ++ ++#include <linux/proc_fs.h> ++ ++#define PESTATUS_ADDR_CLASS 0x800 ++#define PESTATUS_ADDR_TMU 0x80 ++#define PESTATUS_ADDR_UTIL 0x0 ++ ++#define TMU_CONTEXT_ADDR 0x3c8 ++#define IPSEC_CNTRS_ADDR 0x840 ++ ++int pfe_sysfs_init(struct pfe *pfe); ++void pfe_sysfs_exit(struct pfe *pfe); ++#endif /* _PFE_SYSFS_H_ */ +--- /dev/null ++++ b/drivers/staging/fsl_ppfe/platform.h +@@ -0,0 +1,25 @@ ++/* ++ * ++ * Copyright (C) 2007 Freescale Semiconductor, Inc. ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation; either version 2 of the License, or ++ * (at your option) any later version. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public License ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ++ */ ++ ++#ifndef _PLATFORM_H_ ++#define _PLATFORM_H_ ++ ++#define virt_to_phys(virt) ((unsigned long)virt) ++ ++#endif /* _PLATFORM_H_ */ +--- a/include/linux/skbuff.h ++++ b/include/linux/skbuff.h +@@ -858,6 +858,17 @@ static inline struct sk_buff *alloc_skb_ + return __alloc_skb(size, priority, SKB_ALLOC_FCLONE, NUMA_NO_NODE); + } + ++extern struct sk_buff *__alloc_skb_header(unsigned int size, void *data, ++ gfp_t gfp_mask, ++ int fclone, ++ int node); ++static inline struct sk_buff *alloc_skb_header(unsigned int size, ++ u8 *data, ++ gfp_t priority) ++{ ++ return __alloc_skb_header(size, data, priority, 0, -1); ++} ++ + struct sk_buff *__alloc_skb_head(gfp_t priority, int node); + static inline struct sk_buff *alloc_skb_head(gfp_t priority) + { +--- a/net/core/skbuff.c ++++ b/net/core/skbuff.c +@@ -283,6 +283,90 @@ nodata: + EXPORT_SYMBOL(__alloc_skb); + + /** ++ * __alloc_skb_header - allocate a network buffer ++ * @size: size to allocate ++ * @gfp_mask: allocation mask ++ * @fclone: allocate from fclone cache instead of head cache ++ * and allocate a cloned (child) skb ++ * ++ * Allocate a new &sk_buff. The returned buffer has no headroom and a ++ * tail room of size bytes. The object has a reference count of one. ++ * The return is the buffer. On a failure the return is %NULL. ++ * ++ * Buffers may only be allocated from interrupts using a @gfp_mask of ++ * %GFP_ATOMIC. ++ */ ++struct sk_buff *__alloc_skb_header(unsigned int size, void *data, ++ gfp_t gfp_mask, int fclone, int node) ++{ ++ struct kmem_cache *cache; ++ struct skb_shared_info *shinfo; ++ struct sk_buff *skb; ++ ++ cache = fclone ? skbuff_fclone_cache : skbuff_head_cache; ++ ++ if (size <= SKB_DATA_ALIGN(sizeof(struct skb_shared_info))) { ++ skb = NULL; ++ goto out; ++ } ++ ++ /* Get the HEAD */ ++ skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node); ++ if (!skb) ++ goto out; ++ prefetchw(skb); ++ ++ /* kmalloc might give us more room than requested. ++ * Put skb_shared_info exactly at the end of allocated zone, ++ * to allow max possible filling before reallocation. ++ */ ++ size = SKB_WITH_OVERHEAD(ksize(data)); ++ prefetchw(data + size); ++ ++ /* Only clear those fields we need to clear, not those that we will ++ * actually initialise below. Hence, don't put any more fields after ++ * the tail pointer in struct sk_buff! ++ */ ++ memset(skb, 0, offsetof(struct sk_buff, tail)); ++ /* Account for allocated memory : skb + skb->head */ ++ skb->truesize = SKB_TRUESIZE(size); ++ atomic_set(&skb->users, 1); ++ skb->head = data; ++ skb->data = data; ++ skb_reset_tail_pointer(skb); ++ skb->end = skb->tail + size; ++#ifdef NET_SKBUFF_DATA_USES_OFFSET ++ skb->mac_header = ~0U; ++#endif ++ ++#if defined(CONFIG_COMCERTO_CUSTOM_SKB_LAYOUT) ++ skb->mspd_data = NULL; ++ skb->mspd_len = 0; ++#endif ++ ++ /* make sure we initialize shinfo sequentially */ ++ shinfo = skb_shinfo(skb); ++ memset(shinfo, 0, offsetof(struct skb_shared_info, dataref)); ++ atomic_set(&shinfo->dataref, 1); ++ kmemcheck_annotate_variable(shinfo->destructor_arg); ++ ++ if (fclone) { ++ struct sk_buff *child = skb + 1; ++ atomic_t *fclone_ref = (atomic_t *)(child + 1); ++ ++ kmemcheck_annotate_bitfield(child, flags1); ++ kmemcheck_annotate_bitfield(child, flags2); ++ skb->fclone = SKB_FCLONE_ORIG; ++ atomic_set(fclone_ref, 1); ++ ++ child->fclone = SKB_FCLONE_UNAVAILABLE; ++ } ++out: ++ return skb; ++} ++EXPORT_SYMBOL(__alloc_skb_header); ++ ++/** + * __build_skb - build a network buffer + * @data: data buffer provided by caller + * @frag_size: size of data, or 0 if head was kmalloced diff --git a/target/linux/layerscape/patches-4.4/8073-ls1012a-added-clock-configuration.patch b/target/linux/layerscape/patches-4.4/8073-ls1012a-added-clock-configuration.patch new file mode 100644 index 0000000..3936c80 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/8073-ls1012a-added-clock-configuration.patch @@ -0,0 +1,61 @@ +From 9112596c3c7b7b8b1eded3323765fa711dc58e74 Mon Sep 17 00:00:00 2001 +From: Tang Yuantian <Yuantian.Tang@nxp.com> +Date: Thu, 25 Aug 2016 10:38:28 +0800 +Subject: [PATCH 073/113] ls1012a: added clock configuration + +commit c9c11181191938b77bfd61e5094a63955cf711fd +[context adjustment] +[don't apply fsl-ls1012a.dtsi] + +Currently ls1012a used the clock configuration of ls1043a's. +But there is a little different between them. This patch added +ls1012a its own clock configuration. + +Signed-off-by: Tang Yuantian <yuantian.tang@nxp.com> +Integrated-by: Zhao Qiang <qiang.zhao@nxp.com> +--- + drivers/clk/clk-qoriq.c | 19 +++++++++++++++++++ + 1 file changed, 19 insertions(+) + +--- a/drivers/clk/clk-qoriq.c ++++ b/drivers/clk/clk-qoriq.c +@@ -195,6 +195,14 @@ static const struct clockgen_muxinfo t10 + } + }; + ++static const struct clockgen_muxinfo ls1012a_cmux = { ++ { ++ [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 }, ++ {}, ++ [2] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 }, ++ } ++}; ++ + static const struct clockgen_muxinfo t1040_cmux = { + { + [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 }, +@@ -475,6 +483,16 @@ static const struct clockgen_chipinfo ch + .pll_mask = 0x03, + }, + { ++ .compat = "fsl,ls1012a-clockgen", ++ .cmux_groups = { ++ &ls1012a_cmux ++ }, ++ .cmux_to_group = { ++ 0, -1 ++ }, ++ .pll_mask = 0x03, ++ }, ++ { + .compat = "fsl,ls1043a-clockgen", + .init_periph = t2080_init_periph, + .cmux_groups = { +@@ -1268,6 +1286,7 @@ CLK_OF_DECLARE(qoriq_clockgen_2, "fsl,qo + CLK_OF_DECLARE(qoriq_clockgen_ls1021a, "fsl,ls1021a-clockgen", clockgen_init); + CLK_OF_DECLARE(qoriq_clockgen_ls1043a, "fsl,ls1043a-clockgen", clockgen_init); + CLK_OF_DECLARE(qoriq_clockgen_ls2080a, "fsl,ls2080a-clockgen", clockgen_init); ++CLK_OF_DECLARE(qoriq_clockgen_ls1012a, "fsl,ls1012a-clockgen", clockgen_init); + + /* Legacy nodes */ + CLK_OF_DECLARE(qoriq_sysclk_1, "fsl,qoriq-sysclk-1.0", sysclk_init); diff --git a/target/linux/layerscape/patches-4.4/8114-drivers-PCIE-enable-for-Linux.patch b/target/linux/layerscape/patches-4.4/8114-drivers-PCIE-enable-for-Linux.patch new file mode 100644 index 0000000..61c85a1 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/8114-drivers-PCIE-enable-for-Linux.patch @@ -0,0 +1,49 @@ +From cfe7a6abd3d7e9ffeed8230847bbe2f680757305 Mon Sep 17 00:00:00 2001 +From: Pratiyush Mohan Srivastava <pratiyush.srivastava@nxp.com> +Date: Sun, 24 Apr 2016 23:43:19 +0530 +Subject: [PATCH 114/123] drivers: PCIE enable for Linux + +[This patch from sdk release, just context adjustment] +Add support for PCIE for LS1012A in kernel + +Signed-off-by: Pratiyush Mohan Srivastava <pratiyush.srivastava@nxp.com> +Signed-off-by: Prabhakar Kushwaha <prabhakar.kushwaha@nxp.com> +Integrated-by: Jiang Yutang <yutang.jiang@nxp.com> +--- + drivers/irqchip/irq-ls-scfg-msi.c | 1 + + drivers/pci/host/pci-layerscape.c | 7 +++++++ + 2 files changed, 8 insertions(+) + +--- a/drivers/irqchip/irq-ls-scfg-msi.c ++++ b/drivers/irqchip/irq-ls-scfg-msi.c +@@ -219,6 +219,7 @@ static int ls_scfg_msi_remove(struct pla + } + + static const struct of_device_id ls_scfg_msi_id[] = { ++ { .compatible = "fsl,ls1012a-msi", }, + { .compatible = "fsl,1s1021a-msi", }, + { .compatible = "fsl,1s1043a-msi", }, + {}, +--- a/drivers/pci/host/pci-layerscape.c ++++ b/drivers/pci/host/pci-layerscape.c +@@ -192,6 +192,12 @@ static struct ls_pcie_drvdata ls1021_drv + .ops = &ls1021_pcie_host_ops, + }; + ++static struct ls_pcie_drvdata ls1012_drvdata = { ++ .lut_offset = 0xC0000, ++ .ltssm_shift = 24, ++ .ops = &ls_pcie_host_ops, ++}; ++ + static struct ls_pcie_drvdata ls1043_drvdata = { + .lut_offset = 0x10000, + .ltssm_shift = 24, +@@ -205,6 +211,7 @@ static struct ls_pcie_drvdata ls2080_drv + }; + + static const struct of_device_id ls_pcie_of_match[] = { ++ { .compatible = "fsl,ls1012a-pcie", .data = &ls1012_drvdata }, + { .compatible = "fsl,ls1021a-pcie", .data = &ls1021_drvdata }, + { .compatible = "fsl,ls1043a-pcie", .data = &ls1043_drvdata }, + { .compatible = "fsl,ls2080a-pcie", .data = &ls2080_drvdata }, diff --git a/target/linux/layerscape/patches-4.4/8115-PCI-layerscape-call-dw_pcie_setup_rc-in-host-initial.patch b/target/linux/layerscape/patches-4.4/8115-PCI-layerscape-call-dw_pcie_setup_rc-in-host-initial.patch new file mode 100644 index 0000000..cf67e59 --- /dev/null +++ b/target/linux/layerscape/patches-4.4/8115-PCI-layerscape-call-dw_pcie_setup_rc-in-host-initial.patch @@ -0,0 +1,30 @@ +From 5d661761fd2354e4f976c83143a1bf7c1ecad766 Mon Sep 17 00:00:00 2001 +From: Yangbo Lu <yangbo.lu@nxp.com> +Date: Tue, 20 Sep 2016 16:07:10 +0800 +Subject: [PATCH 115/115] PCI: layerscape: call dw_pcie_setup_rc() in host + initialization + +A previous patch moved some Root Complex programming from +dw_pcie_host_init() to dw_pcie_setup_rc() where it belongs, +while the pci-layerscape driver didn't call dw_pcie_setup_rc() +anywhere after that. This patch is to add dw_pcie_setup_rc() +calling in layerscape host initialization to fix ls1012a pci +issue caused by that patch. + +Fixes: c49b76f3c613("PCI: designware: Move Root Complex setup + code to dw_pcie_setup_rc()") +Signed-off-by: Yangbo Lu <yangbo.lu@nxp.com> +--- + drivers/pci/host/pci-layerscape.c | 1 + + 1 file changed, 1 insertion(+) + +--- a/drivers/pci/host/pci-layerscape.c ++++ b/drivers/pci/host/pci-layerscape.c +@@ -148,6 +148,7 @@ static void ls_pcie_host_init(struct pci + { + struct ls_pcie *pcie = to_ls_pcie(pp); + ++ dw_pcie_setup_rc(pp); + iowrite32(1, pcie->dbi + PCIE_DBI_RO_WR_EN); + ls_pcie_fix_class(pcie); + ls_pcie_clear_multifunction(pcie); |