| /* |
| * Blackfin CPLB initialization |
| * |
| * Copyright 2007-2009 Analog Devices Inc. |
| * |
| * Licensed under the GPL-2 or later. |
| */ |
| |
| #include <linux/module.h> |
| |
| #include <asm/blackfin.h> |
| #include <asm/cacheflush.h> |
| #include <asm/cplb.h> |
| #include <asm/cplbinit.h> |
| #include <asm/mem_map.h> |
| |
| struct cplb_entry icplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR; |
| struct cplb_entry dcplb_tbl[NR_CPUS][MAX_CPLBS] PDT_ATTR; |
| |
| int first_switched_icplb PDT_ATTR; |
| int first_switched_dcplb PDT_ATTR; |
| |
| struct cplb_boundary dcplb_bounds[9] PDT_ATTR; |
| struct cplb_boundary icplb_bounds[9] PDT_ATTR; |
| |
| int icplb_nr_bounds PDT_ATTR; |
| int dcplb_nr_bounds PDT_ATTR; |
| |
| void __init generate_cplb_tables_cpu(unsigned int cpu) |
| { |
| int i_d, i_i; |
| unsigned long addr; |
| |
| struct cplb_entry *d_tbl = dcplb_tbl[cpu]; |
| struct cplb_entry *i_tbl = icplb_tbl[cpu]; |
| |
| printk(KERN_INFO "NOMPU: setting up cplb tables\n"); |
| |
| i_d = i_i = 0; |
| |
| #ifdef CONFIG_DEBUG_HUNT_FOR_ZERO |
| /* Set up the zero page. */ |
| d_tbl[i_d].addr = 0; |
| d_tbl[i_d++].data = SDRAM_OOPS | PAGE_SIZE_1KB; |
| i_tbl[i_i].addr = 0; |
| i_tbl[i_i++].data = SDRAM_OOPS | PAGE_SIZE_1KB; |
| #endif |
| |
| /* Cover kernel memory with 4M pages. */ |
| addr = 0; |
| |
| for (; addr < memory_start; addr += 4 * 1024 * 1024) { |
| d_tbl[i_d].addr = addr; |
| d_tbl[i_d++].data = SDRAM_DGENERIC | PAGE_SIZE_4MB; |
| i_tbl[i_i].addr = addr; |
| i_tbl[i_i++].data = SDRAM_IGENERIC | PAGE_SIZE_4MB; |
| } |
| |
| /* Cover L1 memory. One 4M area for code and data each is enough. */ |
| if (cpu == 0) { |
| if (L1_DATA_A_LENGTH || L1_DATA_B_LENGTH) { |
| d_tbl[i_d].addr = L1_DATA_A_START; |
| d_tbl[i_d++].data = L1_DMEMORY | PAGE_SIZE_4MB; |
| } |
| i_tbl[i_i].addr = L1_CODE_START; |
| i_tbl[i_i++].data = L1_IMEMORY | PAGE_SIZE_4MB; |
| } |
| #ifdef CONFIG_SMP |
| else { |
| if (L1_DATA_A_LENGTH || L1_DATA_B_LENGTH) { |
| d_tbl[i_d].addr = COREB_L1_DATA_A_START; |
| d_tbl[i_d++].data = L1_DMEMORY | PAGE_SIZE_4MB; |
| } |
| i_tbl[i_i].addr = COREB_L1_CODE_START; |
| i_tbl[i_i++].data = L1_IMEMORY | PAGE_SIZE_4MB; |
| } |
| #endif |
| first_switched_dcplb = i_d; |
| first_switched_icplb = i_i; |
| |
| BUG_ON(first_switched_dcplb > MAX_CPLBS); |
| BUG_ON(first_switched_icplb > MAX_CPLBS); |
| |
| while (i_d < MAX_CPLBS) |
| d_tbl[i_d++].data = 0; |
| while (i_i < MAX_CPLBS) |
| i_tbl[i_i++].data = 0; |
| } |
| |
| void __init generate_cplb_tables_all(void) |
| { |
| unsigned long uncached_end; |
| int i_d, i_i; |
| |
| i_d = 0; |
| /* Normal RAM, including MTD FS. */ |
| #ifdef CONFIG_MTD_UCLINUX |
| uncached_end = memory_mtd_start + mtd_size; |
| #else |
| uncached_end = memory_end; |
| #endif |
| /* |
| * if DMA uncached is less than 1MB, mark the 1MB chunk as uncached |
| * so that we don't have to use 4kB pages and cause CPLB thrashing |
| */ |
| if ((DMA_UNCACHED_REGION >= 1 * 1024 * 1024) || !DMA_UNCACHED_REGION || |
| ((_ramend - uncached_end) >= 1 * 1024 * 1024)) |
| dcplb_bounds[i_d].eaddr = uncached_end; |
| else |
| dcplb_bounds[i_d].eaddr = uncached_end & ~(1 * 1024 * 1024); |
| dcplb_bounds[i_d++].data = SDRAM_DGENERIC; |
| /* DMA uncached region. */ |
| if (DMA_UNCACHED_REGION) { |
| dcplb_bounds[i_d].eaddr = _ramend; |
| dcplb_bounds[i_d++].data = SDRAM_DNON_CHBL; |
| } |
| if (_ramend != physical_mem_end) { |
| /* Reserved memory. */ |
| dcplb_bounds[i_d].eaddr = physical_mem_end; |
| dcplb_bounds[i_d++].data = (reserved_mem_dcache_on ? |
| SDRAM_DGENERIC : SDRAM_DNON_CHBL); |
| } |
| /* Addressing hole up to the async bank. */ |
| dcplb_bounds[i_d].eaddr = ASYNC_BANK0_BASE; |
| dcplb_bounds[i_d++].data = 0; |
| /* ASYNC banks. */ |
| dcplb_bounds[i_d].eaddr = ASYNC_BANK3_BASE + ASYNC_BANK3_SIZE; |
| dcplb_bounds[i_d++].data = SDRAM_EBIU; |
| /* Addressing hole up to BootROM. */ |
| dcplb_bounds[i_d].eaddr = BOOT_ROM_START; |
| dcplb_bounds[i_d++].data = 0; |
| /* BootROM -- largest one should be less than 1 meg. */ |
| dcplb_bounds[i_d].eaddr = BOOT_ROM_START + (1 * 1024 * 1024); |
| dcplb_bounds[i_d++].data = SDRAM_DGENERIC; |
| if (L2_LENGTH) { |
| /* Addressing hole up to L2 SRAM. */ |
| dcplb_bounds[i_d].eaddr = L2_START; |
| dcplb_bounds[i_d++].data = 0; |
| /* L2 SRAM. */ |
| dcplb_bounds[i_d].eaddr = L2_START + L2_LENGTH; |
| dcplb_bounds[i_d++].data = L2_DMEMORY; |
| } |
| dcplb_nr_bounds = i_d; |
| BUG_ON(dcplb_nr_bounds > ARRAY_SIZE(dcplb_bounds)); |
| |
| i_i = 0; |
| /* Normal RAM, including MTD FS. */ |
| icplb_bounds[i_i].eaddr = uncached_end; |
| icplb_bounds[i_i++].data = SDRAM_IGENERIC; |
| if (_ramend != physical_mem_end) { |
| /* DMA uncached region. */ |
| if (DMA_UNCACHED_REGION) { |
| /* Normally this hole is caught by the async below. */ |
| icplb_bounds[i_i].eaddr = _ramend; |
| icplb_bounds[i_i++].data = 0; |
| } |
| /* Reserved memory. */ |
| icplb_bounds[i_i].eaddr = physical_mem_end; |
| icplb_bounds[i_i++].data = (reserved_mem_icache_on ? |
| SDRAM_IGENERIC : SDRAM_INON_CHBL); |
| } |
| /* Addressing hole up to the async bank. */ |
| icplb_bounds[i_i].eaddr = ASYNC_BANK0_BASE; |
| icplb_bounds[i_i++].data = 0; |
| /* ASYNC banks. */ |
| icplb_bounds[i_i].eaddr = ASYNC_BANK3_BASE + ASYNC_BANK3_SIZE; |
| icplb_bounds[i_i++].data = SDRAM_EBIU; |
| /* Addressing hole up to BootROM. */ |
| icplb_bounds[i_i].eaddr = BOOT_ROM_START; |
| icplb_bounds[i_i++].data = 0; |
| /* BootROM -- largest one should be less than 1 meg. */ |
| icplb_bounds[i_i].eaddr = BOOT_ROM_START + (1 * 1024 * 1024); |
| icplb_bounds[i_i++].data = SDRAM_IGENERIC; |
| |
| if (L2_LENGTH) { |
| /* Addressing hole up to L2 SRAM. */ |
| icplb_bounds[i_i].eaddr = L2_START; |
| icplb_bounds[i_i++].data = 0; |
| /* L2 SRAM. */ |
| icplb_bounds[i_i].eaddr = L2_START + L2_LENGTH; |
| icplb_bounds[i_i++].data = L2_IMEMORY; |
| } |
| icplb_nr_bounds = i_i; |
| BUG_ON(icplb_nr_bounds > ARRAY_SIZE(icplb_bounds)); |
| } |