mmc: core: Reset HPI enabled state during re-init and in case of errors
[linux/fpc-iii.git] / drivers / soc / fsl / qbman / bman_ccsr.c
blob7c3cc968053cd1ddac6230765d6580227e7ee70a
1 /* Copyright (c) 2009 - 2016 Freescale Semiconductor, Inc.
3 * Redistribution and use in source and binary forms, with or without
4 * modification, are permitted provided that the following conditions are met:
5 * * Redistributions of source code must retain the above copyright
6 * notice, this list of conditions and the following disclaimer.
7 * * Redistributions in binary form must reproduce the above copyright
8 * notice, this list of conditions and the following disclaimer in the
9 * documentation and/or other materials provided with the distribution.
10 * * Neither the name of Freescale Semiconductor nor the
11 * names of its contributors may be used to endorse or promote products
12 * derived from this software without specific prior written permission.
14 * ALTERNATIVELY, this software may be distributed under the terms of the
15 * GNU General Public License ("GPL") as published by the Free Software
16 * Foundation, either version 2 of that License or (at your option) any
17 * later version.
19 * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY
20 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
21 * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
22 * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY
23 * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
24 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
25 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
26 * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
27 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
28 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
31 #include "bman_priv.h"
33 u16 bman_ip_rev;
34 EXPORT_SYMBOL(bman_ip_rev);
36 /* Register offsets */
37 #define REG_FBPR_FPC 0x0800
38 #define REG_ECSR 0x0a00
39 #define REG_ECIR 0x0a04
40 #define REG_EADR 0x0a08
41 #define REG_EDATA(n) (0x0a10 + ((n) * 0x04))
42 #define REG_SBEC(n) (0x0a80 + ((n) * 0x04))
43 #define REG_IP_REV_1 0x0bf8
44 #define REG_IP_REV_2 0x0bfc
45 #define REG_FBPR_BARE 0x0c00
46 #define REG_FBPR_BAR 0x0c04
47 #define REG_FBPR_AR 0x0c10
48 #define REG_SRCIDR 0x0d04
49 #define REG_LIODNR 0x0d08
50 #define REG_ERR_ISR 0x0e00
51 #define REG_ERR_IER 0x0e04
52 #define REG_ERR_ISDR 0x0e08
54 /* Used by all error interrupt registers except 'inhibit' */
55 #define BM_EIRQ_IVCI 0x00000010 /* Invalid Command Verb */
56 #define BM_EIRQ_FLWI 0x00000008 /* FBPR Low Watermark */
57 #define BM_EIRQ_MBEI 0x00000004 /* Multi-bit ECC Error */
58 #define BM_EIRQ_SBEI 0x00000002 /* Single-bit ECC Error */
59 #define BM_EIRQ_BSCN 0x00000001 /* pool State Change Notification */
61 struct bman_hwerr_txt {
62 u32 mask;
63 const char *txt;
66 static const struct bman_hwerr_txt bman_hwerr_txts[] = {
67 { BM_EIRQ_IVCI, "Invalid Command Verb" },
68 { BM_EIRQ_FLWI, "FBPR Low Watermark" },
69 { BM_EIRQ_MBEI, "Multi-bit ECC Error" },
70 { BM_EIRQ_SBEI, "Single-bit ECC Error" },
71 { BM_EIRQ_BSCN, "Pool State Change Notification" },
74 /* Only trigger low water mark interrupt once only */
75 #define BMAN_ERRS_TO_DISABLE BM_EIRQ_FLWI
77 /* Pointer to the start of the BMan's CCSR space */
78 static u32 __iomem *bm_ccsr_start;
80 static inline u32 bm_ccsr_in(u32 offset)
82 return ioread32be(bm_ccsr_start + offset/4);
84 static inline void bm_ccsr_out(u32 offset, u32 val)
86 iowrite32be(val, bm_ccsr_start + offset/4);
89 static void bm_get_version(u16 *id, u8 *major, u8 *minor)
91 u32 v = bm_ccsr_in(REG_IP_REV_1);
92 *id = (v >> 16);
93 *major = (v >> 8) & 0xff;
94 *minor = v & 0xff;
97 /* signal transactions for FBPRs with higher priority */
98 #define FBPR_AR_RPRIO_HI BIT(30)
100 static void bm_set_memory(u64 ba, u32 size)
102 u32 exp = ilog2(size);
103 /* choke if size isn't within range */
104 DPAA_ASSERT(size >= 4096 && size <= 1024*1024*1024 &&
105 is_power_of_2(size));
106 /* choke if '[e]ba' has lower-alignment than 'size' */
107 DPAA_ASSERT(!(ba & (size - 1)));
108 bm_ccsr_out(REG_FBPR_BARE, upper_32_bits(ba));
109 bm_ccsr_out(REG_FBPR_BAR, lower_32_bits(ba));
110 bm_ccsr_out(REG_FBPR_AR, exp - 1);
114 * Location and size of BMan private memory
116 * Ideally we would use the DMA API to turn rmem->base into a DMA address
117 * (especially if iommu translations ever get involved). Unfortunately, the
118 * DMA API currently does not allow mapping anything that is not backed with
119 * a struct page.
121 static dma_addr_t fbpr_a;
122 static size_t fbpr_sz;
123 static int __bman_probed;
125 static int bman_fbpr(struct reserved_mem *rmem)
127 fbpr_a = rmem->base;
128 fbpr_sz = rmem->size;
130 WARN_ON(!(fbpr_a && fbpr_sz));
132 return 0;
134 RESERVEDMEM_OF_DECLARE(bman_fbpr, "fsl,bman-fbpr", bman_fbpr);
136 static irqreturn_t bman_isr(int irq, void *ptr)
138 u32 isr_val, ier_val, ecsr_val, isr_mask, i;
139 struct device *dev = ptr;
141 ier_val = bm_ccsr_in(REG_ERR_IER);
142 isr_val = bm_ccsr_in(REG_ERR_ISR);
143 ecsr_val = bm_ccsr_in(REG_ECSR);
144 isr_mask = isr_val & ier_val;
146 if (!isr_mask)
147 return IRQ_NONE;
149 for (i = 0; i < ARRAY_SIZE(bman_hwerr_txts); i++) {
150 if (bman_hwerr_txts[i].mask & isr_mask) {
151 dev_err_ratelimited(dev, "ErrInt: %s\n",
152 bman_hwerr_txts[i].txt);
153 if (bman_hwerr_txts[i].mask & ecsr_val) {
154 /* Re-arm error capture registers */
155 bm_ccsr_out(REG_ECSR, ecsr_val);
157 if (bman_hwerr_txts[i].mask & BMAN_ERRS_TO_DISABLE) {
158 dev_dbg(dev, "Disabling error 0x%x\n",
159 bman_hwerr_txts[i].mask);
160 ier_val &= ~bman_hwerr_txts[i].mask;
161 bm_ccsr_out(REG_ERR_IER, ier_val);
165 bm_ccsr_out(REG_ERR_ISR, isr_val);
167 return IRQ_HANDLED;
170 int bman_is_probed(void)
172 return __bman_probed;
174 EXPORT_SYMBOL_GPL(bman_is_probed);
176 static int fsl_bman_probe(struct platform_device *pdev)
178 int ret, err_irq;
179 struct device *dev = &pdev->dev;
180 struct device_node *node = dev->of_node;
181 struct resource *res;
182 u16 id, bm_pool_cnt;
183 u8 major, minor;
185 __bman_probed = -1;
187 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
188 if (!res) {
189 dev_err(dev, "Can't get %pOF property 'IORESOURCE_MEM'\n",
190 node);
191 return -ENXIO;
193 bm_ccsr_start = devm_ioremap(dev, res->start, resource_size(res));
194 if (!bm_ccsr_start)
195 return -ENXIO;
197 bm_get_version(&id, &major, &minor);
198 if (major == 1 && minor == 0) {
199 bman_ip_rev = BMAN_REV10;
200 bm_pool_cnt = BM_POOL_MAX;
201 } else if (major == 2 && minor == 0) {
202 bman_ip_rev = BMAN_REV20;
203 bm_pool_cnt = 8;
204 } else if (major == 2 && minor == 1) {
205 bman_ip_rev = BMAN_REV21;
206 bm_pool_cnt = BM_POOL_MAX;
207 } else {
208 dev_err(dev, "Unknown Bman version:%04x,%02x,%02x\n",
209 id, major, minor);
210 return -ENODEV;
214 * If FBPR memory wasn't defined using the qbman compatible string
215 * try using the of_reserved_mem_device method
217 if (!fbpr_a) {
218 ret = qbman_init_private_mem(dev, 0, &fbpr_a, &fbpr_sz);
219 if (ret) {
220 dev_err(dev, "qbman_init_private_mem() failed 0x%x\n",
221 ret);
222 return -ENODEV;
226 dev_dbg(dev, "Allocated FBPR 0x%llx 0x%zx\n", fbpr_a, fbpr_sz);
228 bm_set_memory(fbpr_a, fbpr_sz);
230 err_irq = platform_get_irq(pdev, 0);
231 if (err_irq <= 0) {
232 dev_info(dev, "Can't get %pOF IRQ\n", node);
233 return -ENODEV;
235 ret = devm_request_irq(dev, err_irq, bman_isr, IRQF_SHARED, "bman-err",
236 dev);
237 if (ret) {
238 dev_err(dev, "devm_request_irq() failed %d for '%pOF'\n",
239 ret, node);
240 return ret;
242 /* Disable Buffer Pool State Change */
243 bm_ccsr_out(REG_ERR_ISDR, BM_EIRQ_BSCN);
245 * Write-to-clear any stale bits, (eg. starvation being asserted prior
246 * to resource allocation during driver init).
248 bm_ccsr_out(REG_ERR_ISR, 0xffffffff);
249 /* Enable Error Interrupts */
250 bm_ccsr_out(REG_ERR_IER, 0xffffffff);
252 bm_bpalloc = devm_gen_pool_create(dev, 0, -1, "bman-bpalloc");
253 if (IS_ERR(bm_bpalloc)) {
254 ret = PTR_ERR(bm_bpalloc);
255 dev_err(dev, "bman-bpalloc pool init failed (%d)\n", ret);
256 return ret;
259 /* seed BMan resource pool */
260 ret = gen_pool_add(bm_bpalloc, DPAA_GENALLOC_OFF, bm_pool_cnt, -1);
261 if (ret) {
262 dev_err(dev, "Failed to seed BPID range [%d..%d] (%d)\n",
263 0, bm_pool_cnt - 1, ret);
264 return ret;
267 __bman_probed = 1;
269 return 0;
272 static const struct of_device_id fsl_bman_ids[] = {
274 .compatible = "fsl,bman",
279 static struct platform_driver fsl_bman_driver = {
280 .driver = {
281 .name = KBUILD_MODNAME,
282 .of_match_table = fsl_bman_ids,
283 .suppress_bind_attrs = true,
285 .probe = fsl_bman_probe,
288 builtin_platform_driver(fsl_bman_driver);