📄 indy_sc.c
字号:
/* $Id: indy_sc.c,v 1.14 2000/03/25 22:35:07 ralf Exp $ * * indy_sc.c: Indy cache managment functions. * * Copyright (C) 1997 Ralf Baechle (ralf@gnu.org), * derived from r4xx0.c by David S. Miller (dm@engr.sgi.com). */#include <linux/init.h>#include <linux/kernel.h>#include <linux/sched.h>#include <linux/mm.h>#include <asm/bcache.h>#include <asm/sgi/sgi.h>#include <asm/sgi/sgimc.h>#include <asm/page.h>#include <asm/pgtable.h>#include <asm/system.h>#include <asm/bootinfo.h>#include <asm/sgialib.h>#include <asm/mmu_context.h>/* Secondary cache size in bytes, if present. */static unsigned long scache_size;#undef DEBUG_CACHE#define SC_SIZE 0x00080000#define SC_LINE 32#define CI_MASK (SC_SIZE - SC_LINE)#define SC_INDEX(n) ((n) & CI_MASK)static inline void indy_sc_wipe(unsigned long first, unsigned long last){ unsigned long tmp; __asm__ __volatile__(" .set noreorder .set mips3 .set noat mfc0 %2, $12 li $1, 0x80 # Go 64 bit mtc0 $1, $12 dli $1, 0x9000000080000000 or %0, $1 # first line to flush or %1, $1 # last line to flush .set at1: sw $0, 0(%0) bne %0, %1, 1b daddu %0, 32 mtc0 %2, $12 # Back to 32 bit nop; nop; nop; nop; .set mips0 .set reorder" : "=r" (first), "=r" (last), "=&r" (tmp) : "0" (first), "1" (last) : "$1");}static void indy_sc_wback_invalidate(unsigned long addr, unsigned long size){ unsigned long first_line, last_line; unsigned int flags;#ifdef DEBUG_CACHE printk("indy_sc_wback_invalidate[%08lx,%08lx]", addr, size);#endif if (!size) return; /* Which lines to flush? */ first_line = SC_INDEX(addr); last_line = SC_INDEX(addr + size - 1); __save_and_cli(flags); if (first_line <= last_line) { indy_sc_wipe(first_line, last_line); goto out; } indy_sc_wipe(first_line, SC_SIZE - SC_LINE); indy_sc_wipe(0, last_line);out: __restore_flags(flags);}static void indy_sc_enable(void){ unsigned long addr, tmp1, tmp2; /* This is really cool... */#ifdef DEBUG_CACHE printk("Enabling R4600 SCACHE\n");#endif __asm__ __volatile__(" .set push .set noreorder .set mips3 mfc0 %2, $12 nop; nop; nop; nop; li %1, 0x80 mtc0 %1, $12 nop; nop; nop; nop; li %0, 0x1 dsll %0, 31 lui %1, 0x9000 dsll32 %1, 0 or %0, %1, %0 sb $0, 0(%0) mtc0 $0, $12 nop; nop; nop; nop; mtc0 %2, $12 nop; nop; nop; nop; .set pop" : "=r" (tmp1), "=r" (tmp2), "=r" (addr));}static void indy_sc_disable(void){ unsigned long tmp1, tmp2, tmp3;#ifdef DEBUG_CACHE printk("Disabling R4600 SCACHE\n");#endif __asm__ __volatile__(" .set push .set noreorder .set mips3 li %0, 0x1 dsll %0, 31 lui %1, 0x9000 dsll32 %1, 0 or %0, %1, %0 mfc0 %2, $12 nop; nop; nop; nop; li %1, 0x80 mtc0 %1, $12 nop; nop; nop; nop; sh $0, 0(%0) mtc0 $0, $12 nop; nop; nop; nop; mtc0 %2, $12 nop; nop; nop; nop; .set pop" : "=r" (tmp1), "=r" (tmp2), "=r" (tmp3));}static inline int __init indy_sc_probe(void){ volatile unsigned int *cpu_control; unsigned short cmd = 0xc220; unsigned long data = 0; int i, n;#ifdef __MIPSEB__ cpu_control = (volatile unsigned int *) KSEG1ADDR(0x1fa00034);#else cpu_control = (volatile unsigned int *) KSEG1ADDR(0x1fa00030);#endif#define DEASSERT(bit) (*(cpu_control) &= (~(bit)))#define ASSERT(bit) (*(cpu_control) |= (bit))#define DELAY for(n = 0; n < 100000; n++) __asm__ __volatile__("") DEASSERT(SGIMC_EEPROM_PRE); DEASSERT(SGIMC_EEPROM_SDATAO); DEASSERT(SGIMC_EEPROM_SECLOCK); DEASSERT(SGIMC_EEPROM_PRE); DELAY; ASSERT(SGIMC_EEPROM_CSEL); ASSERT(SGIMC_EEPROM_SECLOCK); for(i = 0; i < 11; i++) { if(cmd & (1<<15)) ASSERT(SGIMC_EEPROM_SDATAO); else DEASSERT(SGIMC_EEPROM_SDATAO); DEASSERT(SGIMC_EEPROM_SECLOCK); ASSERT(SGIMC_EEPROM_SECLOCK); cmd <<= 1; } DEASSERT(SGIMC_EEPROM_SDATAO); for(i = 0; i < (sizeof(unsigned short) * 8); i++) { unsigned int tmp; DEASSERT(SGIMC_EEPROM_SECLOCK); DELAY; ASSERT(SGIMC_EEPROM_SECLOCK); DELAY; data <<= 1; tmp = *cpu_control; if(tmp & SGIMC_EEPROM_SDATAI) data |= 1; } DEASSERT(SGIMC_EEPROM_SECLOCK); DEASSERT(SGIMC_EEPROM_CSEL); ASSERT(SGIMC_EEPROM_PRE); ASSERT(SGIMC_EEPROM_SECLOCK); data <<= PAGE_SHIFT; if (data == 0) return 0; scache_size = data; printk("R4600/R5000 SCACHE size %ldK, linesize 32 bytes.\n", scache_size >> 10); return 1;}/* XXX Check with wje if the Indy caches can differenciate between writeback + invalidate and just invalidate. */struct bcache_ops indy_sc_ops = { indy_sc_enable, indy_sc_disable, indy_sc_wback_invalidate, indy_sc_wback_invalidate};void __init indy_sc_init(void){ if (indy_sc_probe()) { indy_sc_enable(); bcops = &indy_sc_ops; }}
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -