| /* SMP global caching code |
| * |
| * Copyright (C) 2010 Red Hat, Inc. All Rights Reserved. |
| * Written by David Howells (dhowells@redhat.com) |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public Licence |
| * as published by the Free Software Foundation; either version |
| * 2 of the Licence, or (at your option) any later version. |
| */ |
| #include <linux/module.h> |
| #include <linux/mm.h> |
| #include <linux/mman.h> |
| #include <linux/threads.h> |
| #include <linux/interrupt.h> |
| #include <asm/page.h> |
| #include <asm/pgtable.h> |
| #include <asm/processor.h> |
| #include <asm/cacheflush.h> |
| #include <asm/io.h> |
| #include <linux/uaccess.h> |
| #include <asm/smp.h> |
| #include "cache-smp.h" |
| |
| DEFINE_SPINLOCK(smp_cache_lock); |
| static unsigned long smp_cache_mask; |
| static unsigned long smp_cache_start; |
| static unsigned long smp_cache_end; |
| static cpumask_t smp_cache_ipi_map; /* Bitmask of cache IPI done CPUs */ |
| |
| /** |
| * smp_cache_interrupt - Handle IPI request to flush caches. |
| * |
| * Handle a request delivered by IPI to flush the current CPU's |
| * caches. The parameters are stored in smp_cache_*. |
| */ |
| void smp_cache_interrupt(void) |
| { |
| unsigned long opr_mask = smp_cache_mask; |
| |
| switch ((enum smp_dcache_ops)(opr_mask & SMP_DCACHE_OP_MASK)) { |
| case SMP_DCACHE_NOP: |
| break; |
| case SMP_DCACHE_INV: |
| mn10300_local_dcache_inv(); |
| break; |
| case SMP_DCACHE_INV_RANGE: |
| mn10300_local_dcache_inv_range(smp_cache_start, smp_cache_end); |
| break; |
| case SMP_DCACHE_FLUSH: |
| mn10300_local_dcache_flush(); |
| break; |
| case SMP_DCACHE_FLUSH_RANGE: |
| mn10300_local_dcache_flush_range(smp_cache_start, |
| smp_cache_end); |
| break; |
| case SMP_DCACHE_FLUSH_INV: |
| mn10300_local_dcache_flush_inv(); |
| break; |
| case SMP_DCACHE_FLUSH_INV_RANGE: |
| mn10300_local_dcache_flush_inv_range(smp_cache_start, |
| smp_cache_end); |
| break; |
| } |
| |
| switch ((enum smp_icache_ops)(opr_mask & SMP_ICACHE_OP_MASK)) { |
| case SMP_ICACHE_NOP: |
| break; |
| case SMP_ICACHE_INV: |
| mn10300_local_icache_inv(); |
| break; |
| case SMP_ICACHE_INV_RANGE: |
| mn10300_local_icache_inv_range(smp_cache_start, smp_cache_end); |
| break; |
| } |
| |
| cpumask_clear_cpu(smp_processor_id(), &smp_cache_ipi_map); |
| } |
| |
| /** |
| * smp_cache_call - Issue an IPI to request the other CPUs flush caches |
| * @opr_mask: Cache operation flags |
| * @start: Start address of request |
| * @end: End address of request |
| * |
| * Send cache flush IPI to other CPUs. This invokes smp_cache_interrupt() |
| * above on those other CPUs and then waits for them to finish. |
| * |
| * The caller must hold smp_cache_lock. |
| */ |
| void smp_cache_call(unsigned long opr_mask, |
| unsigned long start, unsigned long end) |
| { |
| smp_cache_mask = opr_mask; |
| smp_cache_start = start; |
| smp_cache_end = end; |
| cpumask_copy(&smp_cache_ipi_map, cpu_online_mask); |
| cpumask_clear_cpu(smp_processor_id(), &smp_cache_ipi_map); |
| |
| send_IPI_allbutself(FLUSH_CACHE_IPI); |
| |
| while (!cpumask_empty(&smp_cache_ipi_map)) |
| /* nothing. lockup detection does not belong here */ |
| mb(); |
| } |