1 // SPDX-License-Identifier: GPL-2.0
3 * Alchemy PCI host mode support.
5 * Copyright 2001-2003, 2007-2008 MontaVista Software Inc.
6 * Author: MontaVista Software, Inc. <source@mvista.com>
8 * Support for all devices (greater than 16) added by David Gathright.
11 #include <linux/clk.h>
12 #include <linux/export.h>
13 #include <linux/types.h>
14 #include <linux/pci.h>
15 #include <linux/platform_device.h>
16 #include <linux/kernel.h>
17 #include <linux/init.h>
18 #include <linux/syscore_ops.h>
19 #include <linux/vmalloc.h>
21 #include <asm/dma-coherence.h>
22 #include <asm/mach-au1x00/au1000.h>
23 #include <asm/tlbmisc.h>
25 #ifdef CONFIG_PCI_DEBUG
26 #define DBG(x...) printk(KERN_DEBUG x)
28 #define DBG(x...) do {} while (0)
31 #define PCI_ACCESS_READ 0
32 #define PCI_ACCESS_WRITE 1
34 struct alchemy_pci_context
{
35 struct pci_controller alchemy_pci_ctrl
; /* leave as first member! */
36 void __iomem
*regs
; /* ctrl base */
37 /* tools for wired entry for config space access */
38 unsigned long last_elo0
;
39 unsigned long last_elo1
;
41 struct vm_struct
*pci_cfg_vm
;
45 int (*board_map_irq
)(const struct pci_dev
*d
, u8 slot
, u8 pin
);
46 int (*board_pci_idsel
)(unsigned int devsel
, int assert);
49 /* for syscore_ops. There's only one PCI controller on Alchemy chips, so this
50 * should suffice for now.
52 static struct alchemy_pci_context
*__alchemy_pci_ctx
;
55 /* IO/MEM resources for PCI. Keep the memres in sync with fixup_bigphys_addr
56 * in arch/mips/alchemy/common/setup.c
58 static struct resource alchemy_pci_def_memres
= {
59 .start
= ALCHEMY_PCI_MEMWIN_START
,
60 .end
= ALCHEMY_PCI_MEMWIN_END
,
61 .name
= "PCI memory space",
62 .flags
= IORESOURCE_MEM
65 static struct resource alchemy_pci_def_iores
= {
66 .start
= ALCHEMY_PCI_IOWIN_START
,
67 .end
= ALCHEMY_PCI_IOWIN_END
,
68 .name
= "PCI IO space",
69 .flags
= IORESOURCE_IO
72 static void mod_wired_entry(int entry
, unsigned long entrylo0
,
73 unsigned long entrylo1
, unsigned long entryhi
,
74 unsigned long pagemask
)
76 unsigned long old_pagemask
;
77 unsigned long old_ctx
;
79 /* Save old context and create impossible VPN2 value */
80 old_ctx
= read_c0_entryhi() & MIPS_ENTRYHI_ASID
;
81 old_pagemask
= read_c0_pagemask();
82 write_c0_index(entry
);
83 write_c0_pagemask(pagemask
);
84 write_c0_entryhi(entryhi
);
85 write_c0_entrylo0(entrylo0
);
86 write_c0_entrylo1(entrylo1
);
88 write_c0_entryhi(old_ctx
);
89 write_c0_pagemask(old_pagemask
);
92 static void alchemy_pci_wired_entry(struct alchemy_pci_context
*ctx
)
94 ctx
->wired_entry
= read_c0_wired();
95 add_wired_entry(0, 0, (unsigned long)ctx
->pci_cfg_vm
->addr
, PM_4K
);
96 ctx
->last_elo0
= ctx
->last_elo1
= ~0;
99 static int config_access(unsigned char access_type
, struct pci_bus
*bus
,
100 unsigned int dev_fn
, unsigned char where
, u32
*data
)
102 struct alchemy_pci_context
*ctx
= bus
->sysdata
;
103 unsigned int device
= PCI_SLOT(dev_fn
);
104 unsigned int function
= PCI_FUNC(dev_fn
);
105 unsigned long offset
, status
, cfg_base
, flags
, entryLo0
, entryLo1
, r
;
106 int error
= PCIBIOS_SUCCESSFUL
;
113 local_irq_save(flags
);
114 r
= __raw_readl(ctx
->regs
+ PCI_REG_STATCMD
) & 0x0000ffff;
115 r
|= PCI_STATCMD_STATUS(0x2000);
116 __raw_writel(r
, ctx
->regs
+ PCI_REG_STATCMD
);
119 /* Allow board vendors to implement their own off-chip IDSEL.
120 * If it doesn't succeed, may as well bail out at this point.
122 if (ctx
->board_pci_idsel(device
, 1) == 0) {
124 local_irq_restore(flags
);
128 /* Setup the config window */
129 if (bus
->number
== 0)
130 cfg_base
= (1 << device
) << 11;
132 cfg_base
= 0x80000000 | (bus
->number
<< 16) | (device
<< 11);
134 /* Setup the lower bits of the 36-bit address */
135 offset
= (function
<< 8) | (where
& ~0x3);
136 /* Pick up any address that falls below the page mask */
137 offset
|= cfg_base
& ~PAGE_MASK
;
140 cfg_base
= cfg_base
& PAGE_MASK
;
142 /* To improve performance, if the current device is the same as
143 * the last device accessed, we don't touch the TLB.
145 entryLo0
= (6 << 26) | (cfg_base
>> 6) | (2 << 3) | 7;
146 entryLo1
= (6 << 26) | (cfg_base
>> 6) | (0x1000 >> 6) | (2 << 3) | 7;
147 if ((entryLo0
!= ctx
->last_elo0
) || (entryLo1
!= ctx
->last_elo1
)) {
148 mod_wired_entry(ctx
->wired_entry
, entryLo0
, entryLo1
,
149 (unsigned long)ctx
->pci_cfg_vm
->addr
, PM_4K
);
150 ctx
->last_elo0
= entryLo0
;
151 ctx
->last_elo1
= entryLo1
;
154 if (access_type
== PCI_ACCESS_WRITE
)
155 __raw_writel(*data
, ctx
->pci_cfg_vm
->addr
+ offset
);
157 *data
= __raw_readl(ctx
->pci_cfg_vm
->addr
+ offset
);
160 DBG("alchemy-pci: cfg access %d bus %u dev %u at %x dat %x conf %lx\n",
161 access_type
, bus
->number
, device
, where
, *data
, offset
);
163 /* check for errors, master abort */
164 status
= __raw_readl(ctx
->regs
+ PCI_REG_STATCMD
);
165 if (status
& (1 << 29)) {
168 DBG("alchemy-pci: master abort on cfg access %d bus %d dev %d\n",
169 access_type
, bus
->number
, device
);
170 } else if ((status
>> 28) & 0xf) {
171 DBG("alchemy-pci: PCI ERR detected: dev %d, status %lx\n",
172 device
, (status
>> 28) & 0xf);
175 __raw_writel(status
& 0xf000ffff, ctx
->regs
+ PCI_REG_STATCMD
);
181 /* Take away the IDSEL. */
182 (void)ctx
->board_pci_idsel(device
, 0);
184 local_irq_restore(flags
);
188 static int read_config_byte(struct pci_bus
*bus
, unsigned int devfn
,
192 int ret
= config_access(PCI_ACCESS_READ
, bus
, devfn
, where
, &data
);
202 static int read_config_word(struct pci_bus
*bus
, unsigned int devfn
,
206 int ret
= config_access(PCI_ACCESS_READ
, bus
, devfn
, where
, &data
);
210 *val
= data
& 0xffff;
214 static int read_config_dword(struct pci_bus
*bus
, unsigned int devfn
,
217 return config_access(PCI_ACCESS_READ
, bus
, devfn
, where
, val
);
220 static int write_config_byte(struct pci_bus
*bus
, unsigned int devfn
,
225 if (config_access(PCI_ACCESS_READ
, bus
, devfn
, where
, &data
))
228 data
= (data
& ~(0xff << ((where
& 3) << 3))) |
229 (val
<< ((where
& 3) << 3));
231 if (config_access(PCI_ACCESS_WRITE
, bus
, devfn
, where
, &data
))
234 return PCIBIOS_SUCCESSFUL
;
237 static int write_config_word(struct pci_bus
*bus
, unsigned int devfn
,
242 if (config_access(PCI_ACCESS_READ
, bus
, devfn
, where
, &data
))
245 data
= (data
& ~(0xffff << ((where
& 3) << 3))) |
246 (val
<< ((where
& 3) << 3));
248 if (config_access(PCI_ACCESS_WRITE
, bus
, devfn
, where
, &data
))
251 return PCIBIOS_SUCCESSFUL
;
254 static int write_config_dword(struct pci_bus
*bus
, unsigned int devfn
,
257 return config_access(PCI_ACCESS_WRITE
, bus
, devfn
, where
, &val
);
260 static int alchemy_pci_read(struct pci_bus
*bus
, unsigned int devfn
,
261 int where
, int size
, u32
*val
)
266 int rc
= read_config_byte(bus
, devfn
, where
, &_val
);
273 int rc
= read_config_word(bus
, devfn
, where
, &_val
);
279 return read_config_dword(bus
, devfn
, where
, val
);
283 static int alchemy_pci_write(struct pci_bus
*bus
, unsigned int devfn
,
284 int where
, int size
, u32 val
)
288 return write_config_byte(bus
, devfn
, where
, (u8
) val
);
290 return write_config_word(bus
, devfn
, where
, (u16
) val
);
292 return write_config_dword(bus
, devfn
, where
, val
);
296 static struct pci_ops alchemy_pci_ops
= {
297 .read
= alchemy_pci_read
,
298 .write
= alchemy_pci_write
,
301 static int alchemy_pci_def_idsel(unsigned int devsel
, int assert)
303 return 1; /* success */
306 /* save PCI controller register contents. */
307 static int alchemy_pci_suspend(void)
309 struct alchemy_pci_context
*ctx
= __alchemy_pci_ctx
;
313 ctx
->pm
[0] = __raw_readl(ctx
->regs
+ PCI_REG_CMEM
);
314 ctx
->pm
[1] = __raw_readl(ctx
->regs
+ PCI_REG_CONFIG
) & 0x0009ffff;
315 ctx
->pm
[2] = __raw_readl(ctx
->regs
+ PCI_REG_B2BMASK_CCH
);
316 ctx
->pm
[3] = __raw_readl(ctx
->regs
+ PCI_REG_B2BBASE0_VID
);
317 ctx
->pm
[4] = __raw_readl(ctx
->regs
+ PCI_REG_B2BBASE1_SID
);
318 ctx
->pm
[5] = __raw_readl(ctx
->regs
+ PCI_REG_MWMASK_DEV
);
319 ctx
->pm
[6] = __raw_readl(ctx
->regs
+ PCI_REG_MWBASE_REV_CCL
);
320 ctx
->pm
[7] = __raw_readl(ctx
->regs
+ PCI_REG_ID
);
321 ctx
->pm
[8] = __raw_readl(ctx
->regs
+ PCI_REG_CLASSREV
);
322 ctx
->pm
[9] = __raw_readl(ctx
->regs
+ PCI_REG_PARAM
);
323 ctx
->pm
[10] = __raw_readl(ctx
->regs
+ PCI_REG_MBAR
);
324 ctx
->pm
[11] = __raw_readl(ctx
->regs
+ PCI_REG_TIMEOUT
);
329 static void alchemy_pci_resume(void)
331 struct alchemy_pci_context
*ctx
= __alchemy_pci_ctx
;
335 __raw_writel(ctx
->pm
[0], ctx
->regs
+ PCI_REG_CMEM
);
336 __raw_writel(ctx
->pm
[2], ctx
->regs
+ PCI_REG_B2BMASK_CCH
);
337 __raw_writel(ctx
->pm
[3], ctx
->regs
+ PCI_REG_B2BBASE0_VID
);
338 __raw_writel(ctx
->pm
[4], ctx
->regs
+ PCI_REG_B2BBASE1_SID
);
339 __raw_writel(ctx
->pm
[5], ctx
->regs
+ PCI_REG_MWMASK_DEV
);
340 __raw_writel(ctx
->pm
[6], ctx
->regs
+ PCI_REG_MWBASE_REV_CCL
);
341 __raw_writel(ctx
->pm
[7], ctx
->regs
+ PCI_REG_ID
);
342 __raw_writel(ctx
->pm
[8], ctx
->regs
+ PCI_REG_CLASSREV
);
343 __raw_writel(ctx
->pm
[9], ctx
->regs
+ PCI_REG_PARAM
);
344 __raw_writel(ctx
->pm
[10], ctx
->regs
+ PCI_REG_MBAR
);
345 __raw_writel(ctx
->pm
[11], ctx
->regs
+ PCI_REG_TIMEOUT
);
347 __raw_writel(ctx
->pm
[1], ctx
->regs
+ PCI_REG_CONFIG
);
350 /* YAMON on all db1xxx boards wipes the TLB and writes zero to C0_wired
351 * on resume, making it necessary to recreate it as soon as possible.
353 ctx
->wired_entry
= 8191; /* impossibly high value */
354 alchemy_pci_wired_entry(ctx
); /* install it */
357 static struct syscore_ops alchemy_pci_pmops
= {
358 .suspend
= alchemy_pci_suspend
,
359 .resume
= alchemy_pci_resume
,
362 static int alchemy_pci_probe(struct platform_device
*pdev
)
364 struct alchemy_pci_platdata
*pd
= pdev
->dev
.platform_data
;
365 struct alchemy_pci_context
*ctx
;
366 void __iomem
*virt_io
;
372 /* need at least PCI IRQ mapping table */
374 dev_err(&pdev
->dev
, "need platform data for PCI setup\n");
379 ctx
= kzalloc(sizeof(*ctx
), GFP_KERNEL
);
381 dev_err(&pdev
->dev
, "no memory for pcictl context\n");
386 r
= platform_get_resource(pdev
, IORESOURCE_MEM
, 0);
388 dev_err(&pdev
->dev
, "no pcictl ctrl regs resource\n");
393 if (!request_mem_region(r
->start
, resource_size(r
), pdev
->name
)) {
394 dev_err(&pdev
->dev
, "cannot claim pci regs\n");
399 c
= clk_get(&pdev
->dev
, "pci_clko");
401 dev_err(&pdev
->dev
, "unable to find PCI clock\n");
406 ret
= clk_prepare_enable(c
);
408 dev_err(&pdev
->dev
, "cannot enable PCI clock\n");
412 ctx
->regs
= ioremap(r
->start
, resource_size(r
));
414 dev_err(&pdev
->dev
, "cannot map pci regs\n");
419 /* map parts of the PCI IO area */
420 /* REVISIT: if this changes with a newer variant (doubt it) make this
421 * a platform resource.
423 virt_io
= ioremap(AU1500_PCI_IO_PHYS_ADDR
, 0x00100000);
425 dev_err(&pdev
->dev
, "cannot remap pci io space\n");
429 ctx
->alchemy_pci_ctrl
.io_map_base
= (unsigned long)virt_io
;
431 /* Au1500 revisions older than AD have borked coherent PCI */
432 if ((alchemy_get_cputype() == ALCHEMY_CPU_AU1500
) &&
433 (read_c0_prid() < 0x01030202) &&
434 (coherentio
== IO_COHERENCE_DISABLED
)) {
435 val
= __raw_readl(ctx
->regs
+ PCI_REG_CONFIG
);
436 val
|= PCI_CONFIG_NC
;
437 __raw_writel(val
, ctx
->regs
+ PCI_REG_CONFIG
);
439 dev_info(&pdev
->dev
, "non-coherent PCI on Au1500 AA/AB/AC\n");
442 if (pd
->board_map_irq
)
443 ctx
->board_map_irq
= pd
->board_map_irq
;
445 if (pd
->board_pci_idsel
)
446 ctx
->board_pci_idsel
= pd
->board_pci_idsel
;
448 ctx
->board_pci_idsel
= alchemy_pci_def_idsel
;
450 /* fill in relevant pci_controller members */
451 ctx
->alchemy_pci_ctrl
.pci_ops
= &alchemy_pci_ops
;
452 ctx
->alchemy_pci_ctrl
.mem_resource
= &alchemy_pci_def_memres
;
453 ctx
->alchemy_pci_ctrl
.io_resource
= &alchemy_pci_def_iores
;
455 /* we can't ioremap the entire pci config space because it's too large,
456 * nor can we dynamically ioremap it because some drivers use the
457 * PCI config routines from within atomic contex and that becomes a
458 * problem in get_vm_area(). Instead we use one wired TLB entry to
459 * handle all config accesses for all busses.
461 ctx
->pci_cfg_vm
= get_vm_area(0x2000, VM_IOREMAP
);
462 if (!ctx
->pci_cfg_vm
) {
463 dev_err(&pdev
->dev
, "unable to get vm area\n");
467 ctx
->wired_entry
= 8191; /* impossibly high value */
468 alchemy_pci_wired_entry(ctx
); /* install it */
470 set_io_port_base((unsigned long)ctx
->alchemy_pci_ctrl
.io_map_base
);
472 /* board may want to modify bits in the config register, do it now */
473 val
= __raw_readl(ctx
->regs
+ PCI_REG_CONFIG
);
474 val
&= ~pd
->pci_cfg_clr
;
475 val
|= pd
->pci_cfg_set
;
476 val
&= ~PCI_CONFIG_PD
; /* clear disable bit */
477 __raw_writel(val
, ctx
->regs
+ PCI_REG_CONFIG
);
480 __alchemy_pci_ctx
= ctx
;
481 platform_set_drvdata(pdev
, ctx
);
482 register_syscore_ops(&alchemy_pci_pmops
);
483 register_pci_controller(&ctx
->alchemy_pci_ctrl
);
485 dev_info(&pdev
->dev
, "PCI controller at %ld MHz\n",
486 clk_get_rate(c
) / 1000000);
495 clk_disable_unprepare(c
);
499 release_mem_region(r
->start
, resource_size(r
));
506 static struct platform_driver alchemy_pcictl_driver
= {
507 .probe
= alchemy_pci_probe
,
509 .name
= "alchemy-pci",
513 static int __init
alchemy_pci_init(void)
515 /* Au1500/Au1550 have PCI */
516 switch (alchemy_get_cputype()) {
517 case ALCHEMY_CPU_AU1500
:
518 case ALCHEMY_CPU_AU1550
:
519 return platform_driver_register(&alchemy_pcictl_driver
);
523 arch_initcall(alchemy_pci_init
);
526 int pcibios_map_irq(const struct pci_dev
*dev
, u8 slot
, u8 pin
)
528 struct alchemy_pci_context
*ctx
= dev
->sysdata
;
529 if (ctx
&& ctx
->board_map_irq
)
530 return ctx
->board_map_irq(dev
, slot
, pin
);
534 int pcibios_plat_dev_init(struct pci_dev
*dev
)