2 * AMD K7 AGPGART routines.
5 #include <linux/module.h>
7 #include <linux/init.h>
8 #include <linux/agp_backend.h>
9 #include <linux/page-flags.h>
11 #include <linux/slab.h>
14 #define AMD_MMBASE 0x14
15 #define AMD_APSIZE 0xac
16 #define AMD_MODECNTL 0xb0
17 #define AMD_MODECNTL2 0xb2
18 #define AMD_GARTENABLE 0x02 /* In mmio region (16-bit register) */
19 #define AMD_ATTBASE 0x04 /* In mmio region (32-bit register) */
20 #define AMD_TLBFLUSH 0x0c /* In mmio region (32-bit register) */
21 #define AMD_CACHEENTRY 0x10 /* In mmio region (32-bit register) */
23 static struct pci_device_id agp_amdk7_pci_table
[];
27 unsigned long __iomem
*remapped
;
30 static struct _amd_irongate_private
{
31 volatile u8 __iomem
*registers
;
32 struct amd_page_map
**gatt_pages
;
34 } amd_irongate_private
;
36 static int amd_create_page_map(struct amd_page_map
*page_map
)
40 page_map
->real
= (unsigned long *) __get_free_page(GFP_KERNEL
);
41 if (page_map
->real
== NULL
)
45 SetPageReserved(virt_to_page(page_map
->real
));
47 page_map
->remapped
= ioremap_nocache(virt_to_phys(page_map
->real
),
49 if (page_map
->remapped
== NULL
) {
50 ClearPageReserved(virt_to_page(page_map
->real
));
51 free_page((unsigned long) page_map
->real
);
52 page_map
->real
= NULL
;
57 set_memory_uc((unsigned long)page_map
->real
, 1);
58 page_map
->remapped
= page_map
->real
;
61 for (i
= 0; i
< PAGE_SIZE
/ sizeof(unsigned long); i
++) {
62 writel(agp_bridge
->scratch_page
, page_map
->remapped
+i
);
63 readl(page_map
->remapped
+i
); /* PCI Posting. */
69 static void amd_free_page_map(struct amd_page_map
*page_map
)
72 iounmap(page_map
->remapped
);
73 ClearPageReserved(virt_to_page(page_map
->real
));
75 set_memory_wb((unsigned long)page_map
->real
, 1);
77 free_page((unsigned long) page_map
->real
);
80 static void amd_free_gatt_pages(void)
83 struct amd_page_map
**tables
;
84 struct amd_page_map
*entry
;
86 tables
= amd_irongate_private
.gatt_pages
;
87 for (i
= 0; i
< amd_irongate_private
.num_tables
; i
++) {
90 if (entry
->real
!= NULL
)
91 amd_free_page_map(entry
);
96 amd_irongate_private
.gatt_pages
= NULL
;
99 static int amd_create_gatt_pages(int nr_tables
)
101 struct amd_page_map
**tables
;
102 struct amd_page_map
*entry
;
106 tables
= kzalloc((nr_tables
+ 1) * sizeof(struct amd_page_map
*),GFP_KERNEL
);
110 for (i
= 0; i
< nr_tables
; i
++) {
111 entry
= kzalloc(sizeof(struct amd_page_map
), GFP_KERNEL
);
117 retval
= amd_create_page_map(entry
);
121 amd_irongate_private
.num_tables
= i
;
122 amd_irongate_private
.gatt_pages
= tables
;
125 amd_free_gatt_pages();
130 /* Since we don't need contiguous memory we just try
131 * to get the gatt table once
134 #define GET_PAGE_DIR_OFF(addr) (addr >> 22)
135 #define GET_PAGE_DIR_IDX(addr) (GET_PAGE_DIR_OFF(addr) - \
136 GET_PAGE_DIR_OFF(agp_bridge->gart_bus_addr))
137 #define GET_GATT_OFF(addr) ((addr & 0x003ff000) >> 12)
138 #define GET_GATT(addr) (amd_irongate_private.gatt_pages[\
139 GET_PAGE_DIR_IDX(addr)]->remapped)
141 static int amd_create_gatt_table(struct agp_bridge_data
*bridge
)
143 struct aper_size_info_lvl2
*value
;
144 struct amd_page_map page_dir
;
145 unsigned long __iomem
*cur_gatt
;
151 value
= A_SIZE_LVL2(agp_bridge
->current_size
);
152 retval
= amd_create_page_map(&page_dir
);
156 retval
= amd_create_gatt_pages(value
->num_entries
/ 1024);
158 amd_free_page_map(&page_dir
);
162 agp_bridge
->gatt_table_real
= (u32
*)page_dir
.real
;
163 agp_bridge
->gatt_table
= (u32 __iomem
*)page_dir
.remapped
;
164 agp_bridge
->gatt_bus_addr
= virt_to_phys(page_dir
.real
);
166 /* Get the address for the gart region.
167 * This is a bus address even on the alpha, b/c its
168 * used to program the agp master not the cpu
171 pci_read_config_dword(agp_bridge
->dev
, AGP_APBASE
, &temp
);
172 addr
= (temp
& PCI_BASE_ADDRESS_MEM_MASK
);
173 agp_bridge
->gart_bus_addr
= addr
;
175 /* Calculate the agp offset */
176 for (i
= 0; i
< value
->num_entries
/ 1024; i
++, addr
+= 0x00400000) {
177 writel(virt_to_phys(amd_irongate_private
.gatt_pages
[i
]->real
) | 1,
178 page_dir
.remapped
+GET_PAGE_DIR_OFF(addr
));
179 readl(page_dir
.remapped
+GET_PAGE_DIR_OFF(addr
)); /* PCI Posting. */
182 for (i
= 0; i
< value
->num_entries
; i
++) {
183 addr
= (i
* PAGE_SIZE
) + agp_bridge
->gart_bus_addr
;
184 cur_gatt
= GET_GATT(addr
);
185 writel(agp_bridge
->scratch_page
, cur_gatt
+GET_GATT_OFF(addr
));
186 readl(cur_gatt
+GET_GATT_OFF(addr
)); /* PCI Posting. */
192 static int amd_free_gatt_table(struct agp_bridge_data
*bridge
)
194 struct amd_page_map page_dir
;
196 page_dir
.real
= (unsigned long *)agp_bridge
->gatt_table_real
;
197 page_dir
.remapped
= (unsigned long __iomem
*)agp_bridge
->gatt_table
;
199 amd_free_gatt_pages();
200 amd_free_page_map(&page_dir
);
204 static int amd_irongate_fetch_size(void)
208 struct aper_size_info_lvl2
*values
;
210 pci_read_config_dword(agp_bridge
->dev
, AMD_APSIZE
, &temp
);
211 temp
= (temp
& 0x0000000e);
212 values
= A_SIZE_LVL2(agp_bridge
->driver
->aperture_sizes
);
213 for (i
= 0; i
< agp_bridge
->driver
->num_aperture_sizes
; i
++) {
214 if (temp
== values
[i
].size_value
) {
215 agp_bridge
->previous_size
=
216 agp_bridge
->current_size
= (void *) (values
+ i
);
218 agp_bridge
->aperture_size_idx
= i
;
219 return values
[i
].size
;
226 static int amd_irongate_configure(void)
228 struct aper_size_info_lvl2
*current_size
;
232 current_size
= A_SIZE_LVL2(agp_bridge
->current_size
);
234 if (!amd_irongate_private
.registers
) {
235 /* Get the memory mapped registers */
236 pci_read_config_dword(agp_bridge
->dev
, AMD_MMBASE
, &temp
);
237 temp
= (temp
& PCI_BASE_ADDRESS_MEM_MASK
);
238 amd_irongate_private
.registers
= (volatile u8 __iomem
*) ioremap(temp
, 4096);
239 if (!amd_irongate_private
.registers
)
243 /* Write out the address of the gatt table */
244 writel(agp_bridge
->gatt_bus_addr
, amd_irongate_private
.registers
+AMD_ATTBASE
);
245 readl(amd_irongate_private
.registers
+AMD_ATTBASE
); /* PCI Posting. */
247 /* Write the Sync register */
248 pci_write_config_byte(agp_bridge
->dev
, AMD_MODECNTL
, 0x80);
250 /* Set indexing mode */
251 pci_write_config_byte(agp_bridge
->dev
, AMD_MODECNTL2
, 0x00);
253 /* Write the enable register */
254 enable_reg
= readw(amd_irongate_private
.registers
+AMD_GARTENABLE
);
255 enable_reg
= (enable_reg
| 0x0004);
256 writew(enable_reg
, amd_irongate_private
.registers
+AMD_GARTENABLE
);
257 readw(amd_irongate_private
.registers
+AMD_GARTENABLE
); /* PCI Posting. */
259 /* Write out the size register */
260 pci_read_config_dword(agp_bridge
->dev
, AMD_APSIZE
, &temp
);
261 temp
= (((temp
& ~(0x0000000e)) | current_size
->size_value
) | 1);
262 pci_write_config_dword(agp_bridge
->dev
, AMD_APSIZE
, temp
);
265 writel(1, amd_irongate_private
.registers
+AMD_TLBFLUSH
);
266 readl(amd_irongate_private
.registers
+AMD_TLBFLUSH
); /* PCI Posting.*/
270 static void amd_irongate_cleanup(void)
272 struct aper_size_info_lvl2
*previous_size
;
276 previous_size
= A_SIZE_LVL2(agp_bridge
->previous_size
);
278 enable_reg
= readw(amd_irongate_private
.registers
+AMD_GARTENABLE
);
279 enable_reg
= (enable_reg
& ~(0x0004));
280 writew(enable_reg
, amd_irongate_private
.registers
+AMD_GARTENABLE
);
281 readw(amd_irongate_private
.registers
+AMD_GARTENABLE
); /* PCI Posting. */
283 /* Write back the previous size and disable gart translation */
284 pci_read_config_dword(agp_bridge
->dev
, AMD_APSIZE
, &temp
);
285 temp
= ((temp
& ~(0x0000000f)) | previous_size
->size_value
);
286 pci_write_config_dword(agp_bridge
->dev
, AMD_APSIZE
, temp
);
287 iounmap((void __iomem
*) amd_irongate_private
.registers
);
291 * This routine could be implemented by taking the addresses
292 * written to the GATT, and flushing them individually. However
293 * currently it just flushes the whole table. Which is probably
294 * more efficent, since agp_memory blocks can be a large number of
298 static void amd_irongate_tlbflush(struct agp_memory
*temp
)
300 writel(1, amd_irongate_private
.registers
+AMD_TLBFLUSH
);
301 readl(amd_irongate_private
.registers
+AMD_TLBFLUSH
); /* PCI Posting. */
304 static int amd_insert_memory(struct agp_memory
*mem
, off_t pg_start
, int type
)
306 int i
, j
, num_entries
;
307 unsigned long __iomem
*cur_gatt
;
310 num_entries
= A_SIZE_LVL2(agp_bridge
->current_size
)->num_entries
;
312 if (type
!= 0 || mem
->type
!= 0)
315 if ((pg_start
+ mem
->page_count
) > num_entries
)
319 while (j
< (pg_start
+ mem
->page_count
)) {
320 addr
= (j
* PAGE_SIZE
) + agp_bridge
->gart_bus_addr
;
321 cur_gatt
= GET_GATT(addr
);
322 if (!PGE_EMPTY(agp_bridge
, readl(cur_gatt
+GET_GATT_OFF(addr
))))
327 if (!mem
->is_flushed
) {
328 global_cache_flush();
329 mem
->is_flushed
= true;
332 for (i
= 0, j
= pg_start
; i
< mem
->page_count
; i
++, j
++) {
333 addr
= (j
* PAGE_SIZE
) + agp_bridge
->gart_bus_addr
;
334 cur_gatt
= GET_GATT(addr
);
335 writel(agp_generic_mask_memory(agp_bridge
,
336 page_to_phys(mem
->pages
[i
]),
338 cur_gatt
+GET_GATT_OFF(addr
));
339 readl(cur_gatt
+GET_GATT_OFF(addr
)); /* PCI Posting. */
341 amd_irongate_tlbflush(mem
);
345 static int amd_remove_memory(struct agp_memory
*mem
, off_t pg_start
, int type
)
348 unsigned long __iomem
*cur_gatt
;
351 if (type
!= 0 || mem
->type
!= 0)
354 for (i
= pg_start
; i
< (mem
->page_count
+ pg_start
); i
++) {
355 addr
= (i
* PAGE_SIZE
) + agp_bridge
->gart_bus_addr
;
356 cur_gatt
= GET_GATT(addr
);
357 writel(agp_bridge
->scratch_page
, cur_gatt
+GET_GATT_OFF(addr
));
358 readl(cur_gatt
+GET_GATT_OFF(addr
)); /* PCI Posting. */
361 amd_irongate_tlbflush(mem
);
365 static const struct aper_size_info_lvl2 amd_irongate_sizes
[7] =
367 {2048, 524288, 0x0000000c},
368 {1024, 262144, 0x0000000a},
369 {512, 131072, 0x00000008},
370 {256, 65536, 0x00000006},
371 {128, 32768, 0x00000004},
372 {64, 16384, 0x00000002},
373 {32, 8192, 0x00000000}
376 static const struct gatt_mask amd_irongate_masks
[] =
378 {.mask
= 1, .type
= 0}
381 static const struct agp_bridge_driver amd_irongate_driver
= {
382 .owner
= THIS_MODULE
,
383 .aperture_sizes
= amd_irongate_sizes
,
384 .size_type
= LVL2_APER_SIZE
,
385 .num_aperture_sizes
= 7,
386 .needs_scratch_page
= true,
387 .configure
= amd_irongate_configure
,
388 .fetch_size
= amd_irongate_fetch_size
,
389 .cleanup
= amd_irongate_cleanup
,
390 .tlb_flush
= amd_irongate_tlbflush
,
391 .mask_memory
= agp_generic_mask_memory
,
392 .masks
= amd_irongate_masks
,
393 .agp_enable
= agp_generic_enable
,
394 .cache_flush
= global_cache_flush
,
395 .create_gatt_table
= amd_create_gatt_table
,
396 .free_gatt_table
= amd_free_gatt_table
,
397 .insert_memory
= amd_insert_memory
,
398 .remove_memory
= amd_remove_memory
,
399 .alloc_by_type
= agp_generic_alloc_by_type
,
400 .free_by_type
= agp_generic_free_by_type
,
401 .agp_alloc_page
= agp_generic_alloc_page
,
402 .agp_alloc_pages
= agp_generic_alloc_pages
,
403 .agp_destroy_page
= agp_generic_destroy_page
,
404 .agp_destroy_pages
= agp_generic_destroy_pages
,
405 .agp_type_to_mask_type
= agp_generic_type_to_mask_type
,
408 static struct agp_device_ids amd_agp_device_ids
[] __devinitdata
=
411 .device_id
= PCI_DEVICE_ID_AMD_FE_GATE_7006
,
412 .chipset_name
= "Irongate",
415 .device_id
= PCI_DEVICE_ID_AMD_FE_GATE_700E
,
416 .chipset_name
= "761",
419 .device_id
= PCI_DEVICE_ID_AMD_FE_GATE_700C
,
420 .chipset_name
= "760MP",
422 { }, /* dummy final entry, always present */
425 static int __devinit
agp_amdk7_probe(struct pci_dev
*pdev
,
426 const struct pci_device_id
*ent
)
428 struct agp_bridge_data
*bridge
;
432 cap_ptr
= pci_find_capability(pdev
, PCI_CAP_ID_AGP
);
436 j
= ent
- agp_amdk7_pci_table
;
437 dev_info(&pdev
->dev
, "AMD %s chipset\n",
438 amd_agp_device_ids
[j
].chipset_name
);
440 bridge
= agp_alloc_bridge();
444 bridge
->driver
= &amd_irongate_driver
;
445 bridge
->dev_private_data
= &amd_irongate_private
,
447 bridge
->capndx
= cap_ptr
;
449 /* 751 Errata (22564_B-1.PDF)
450 erratum 20: strobe glitch with Nvidia NV10 GeForce cards.
451 system controller may experience noise due to strong drive strengths
453 if (agp_bridge
->dev
->device
== PCI_DEVICE_ID_AMD_FE_GATE_7006
) {
454 struct pci_dev
*gfxcard
=NULL
;
458 gfxcard
= pci_get_class(PCI_CLASS_DISPLAY_VGA
<<8, gfxcard
);
460 dev_info(&pdev
->dev
, "no AGP VGA controller\n");
463 cap_ptr
= pci_find_capability(gfxcard
, PCI_CAP_ID_AGP
);
466 /* With so many variants of NVidia cards, it's simpler just
467 to blacklist them all, and then whitelist them as needed
468 (if necessary at all). */
469 if (gfxcard
->vendor
== PCI_VENDOR_ID_NVIDIA
) {
470 agp_bridge
->flags
|= AGP_ERRATA_1X
;
471 dev_info(&pdev
->dev
, "AMD 751 chipset with NVidia GeForce; forcing 1X due to errata\n");
473 pci_dev_put(gfxcard
);
476 /* 761 Errata (23613_F.pdf)
477 * Revisions B0/B1 were a disaster.
478 * erratum 44: SYSCLK/AGPCLK skew causes 2X failures -- Force mode to 1X
479 * erratum 45: Timing problem prevents fast writes -- Disable fast write.
480 * erratum 46: Setup violation on AGP SBA pins - Disable side band addressing.
481 * With this lot disabled, we should prevent lockups. */
482 if (agp_bridge
->dev
->device
== PCI_DEVICE_ID_AMD_FE_GATE_700E
) {
483 if (pdev
->revision
== 0x10 || pdev
->revision
== 0x11) {
484 agp_bridge
->flags
= AGP_ERRATA_FASTWRITES
;
485 agp_bridge
->flags
|= AGP_ERRATA_SBA
;
486 agp_bridge
->flags
|= AGP_ERRATA_1X
;
487 dev_info(&pdev
->dev
, "AMD 761 chipset with errata; disabling AGP fast writes & SBA and forcing to 1X\n");
491 /* Fill in the mode register */
492 pci_read_config_dword(pdev
,
493 bridge
->capndx
+PCI_AGP_STATUS
,
496 pci_set_drvdata(pdev
, bridge
);
497 return agp_add_bridge(bridge
);
500 static void __devexit
agp_amdk7_remove(struct pci_dev
*pdev
)
502 struct agp_bridge_data
*bridge
= pci_get_drvdata(pdev
);
504 agp_remove_bridge(bridge
);
505 agp_put_bridge(bridge
);
510 static int agp_amdk7_suspend(struct pci_dev
*pdev
, pm_message_t state
)
512 pci_save_state(pdev
);
513 pci_set_power_state(pdev
, pci_choose_state(pdev
, state
));
518 static int agp_amdk7_resume(struct pci_dev
*pdev
)
520 pci_set_power_state(pdev
, PCI_D0
);
521 pci_restore_state(pdev
);
523 return amd_irongate_driver
.configure();
526 #endif /* CONFIG_PM */
528 /* must be the same order as name table above */
529 static struct pci_device_id agp_amdk7_pci_table
[] = {
531 .class = (PCI_CLASS_BRIDGE_HOST
<< 8),
533 .vendor
= PCI_VENDOR_ID_AMD
,
534 .device
= PCI_DEVICE_ID_AMD_FE_GATE_7006
,
535 .subvendor
= PCI_ANY_ID
,
536 .subdevice
= PCI_ANY_ID
,
539 .class = (PCI_CLASS_BRIDGE_HOST
<< 8),
541 .vendor
= PCI_VENDOR_ID_AMD
,
542 .device
= PCI_DEVICE_ID_AMD_FE_GATE_700E
,
543 .subvendor
= PCI_ANY_ID
,
544 .subdevice
= PCI_ANY_ID
,
547 .class = (PCI_CLASS_BRIDGE_HOST
<< 8),
549 .vendor
= PCI_VENDOR_ID_AMD
,
550 .device
= PCI_DEVICE_ID_AMD_FE_GATE_700C
,
551 .subvendor
= PCI_ANY_ID
,
552 .subdevice
= PCI_ANY_ID
,
557 MODULE_DEVICE_TABLE(pci
, agp_amdk7_pci_table
);
559 static struct pci_driver agp_amdk7_pci_driver
= {
560 .name
= "agpgart-amdk7",
561 .id_table
= agp_amdk7_pci_table
,
562 .probe
= agp_amdk7_probe
,
563 .remove
= agp_amdk7_remove
,
565 .suspend
= agp_amdk7_suspend
,
566 .resume
= agp_amdk7_resume
,
570 static int __init
agp_amdk7_init(void)
574 return pci_register_driver(&agp_amdk7_pci_driver
);
577 static void __exit
agp_amdk7_cleanup(void)
579 pci_unregister_driver(&agp_amdk7_pci_driver
);
582 module_init(agp_amdk7_init
);
583 module_exit(agp_amdk7_cleanup
);
585 MODULE_LICENSE("GPL and additional rights");