⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 proc-arm922.s

📁 嵌入式系统设计与实验教材二源码linux内核移植与编译
💻 S
📖 第 1 页 / 共 2 页
字号:
/* *  linux/arch/arm/mm/arm922.S: MMU functions for ARM922 * *  Copyright (C) 1999,2000 ARM Limited *  Copyright (C) 2000 Deep Blue Solutions Ltd. *  Copyright (C) 2001 Altera Corporation * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA * * * These are the low level assembler for performing cache and TLB * functions on the arm922. */#include <linux/linkage.h>#include <linux/config.h>#include <asm/assembler.h>#include <asm/constants.h>#include <asm/procinfo.h>#include <asm/hardware.h>/* * This is the maximum size of an area which will be invalidated * using the single invalidate entry instructions.  Anything larger * than this, and we go for the whole cache. * * This value should be chosen such that we choose the cheapest * alternative. */#define MAX_AREA_SIZE	8192/* * the cache line size of the I and D cache */#define DCACHELINESIZE	32#define ICACHELINESIZE	32/* * and the page size */#define PAGESIZE	4096	.text/* * cpu_arm922_data_abort() * * obtain information about current aborted instruction * * r0 = address of aborted instruction * * Returns: *  r0 = address of abort *  r1 != 0 if writing *  r3 = FSR */	.align	5ENTRY(cpu_arm922_data_abort)	ldr	r1, [r0]			@ read aborted instruction	mrc	p15, 0, r0, c6, c0, 0		@ get FAR	tst	r1, r1, lsr #21			@ C = bit 20	mrc	p15, 0, r3, c5, c0, 0		@ get FSR	sbc	r1, r1, r1			@ r1 = C - 1	and	r3, r3, #255	mov	pc, lr	/* * cpu_arm922_check_bugs() */ENTRY(cpu_arm922_check_bugs)	mrs	ip, cpsr	bic	ip, ip, #F_BIT	msr	cpsr, ip	mov	pc, lr/* * cpu_arm922_proc_init() */ENTRY(cpu_arm922_proc_init)	mov	pc, lr/* * cpu_arm922_proc_fin() */ENTRY(cpu_arm922_proc_fin)	stmfd	sp!, {lr}	mov	ip, #F_BIT | I_BIT | SVC_MODE	msr	cpsr_c, ip	bl	cpu_arm922_cache_clean_invalidate_all	mrc	p15, 0, r0, c1, c0, 0		@ ctrl register	bic	r0, r0, #0x1000			@ ...i............	bic	r0, r0, #0x000e			@ ............wca.	mcr	p15, 0, r0, c1, c0, 0		@ disable caches	ldmfd	sp!, {pc}/* * cpu_arm922_reset(loc) * * Perform a soft reset of the system.  Put the CPU into the * same state as it would be if it had been reset, and branch * to what would be the reset vector. * * loc: location to jump to for soft reset */	.align	5ENTRY(cpu_arm922_reset)	mov	ip, #0	mcr	p15, 0, ip, c7, c7, 0		@ invalidate I,D caches	mcr	p15, 0, ip, c7, c10, 4		@ drain WB	mcr	p15, 0, ip, c8, c7, 0		@ invalidate I & D TLBs	mrc	p15, 0, ip, c1, c0, 0		@ ctrl register	bic	ip, ip, #0x000f			@ ............wcam	bic	ip, ip, #0x1100			@ ...i...s........	mcr	p15, 0, ip, c1, c0, 0		@ ctrl register	mov	pc, r0/* * cpu_arm922_do_idle() */	.align	5ENTRY(cpu_arm922_do_idle)#if defined(CONFIG_CPU_ARM922_CPU_IDLE)	mcr	p15, 0, r0, c7, c0, 4		@ Wait for interrupt#endif	mov	pc, lr/* ================================= CACHE ================================ *//* * cpu_arm922_cache_clean_invalidate_all() * * clean and invalidate all cache lines * * Note: *  1. we should preserve r0 at all times */	.align	5ENTRY(cpu_arm922_cache_clean_invalidate_all)	mov	r2, #1cpu_arm922_cache_clean_invalidate_all_r2:	mov	ip, #0#ifdef CONFIG_CPU_ARM922_WRITETHROUGH	mcr	p15, 0, ip, c7, c6, 0		@ invalidate D cache#else/* * 'Clean & Invalidate whole DCache' * Re-written to use Index Ops. * Uses registers r1, r3 and ip */	mov	r1, #3 << 5			@ 4 segments1:	orr	r3, r1, #63 << 26		@ 64 entries2:	mcr	p15, 0, r3, c7, c14, 2		@ clean & invalidate D index	subs	r3, r3, #1 << 26	bcs	2b				@ entries 63 to 0	subs	r1, r1, #1 << 5	bcs	1b				@ segments 7 to 0#endif	teq	r2, #0	mcrne	p15, 0, ip, c7, c5, 0		@ invalidate I cache	mcr	p15, 0, ip, c7, c10, 4		@ drain WB	mov	pc, lr/* * cpu_arm922_cache_clean_invalidate_range(start, end, flags) * * clean and invalidate all cache lines associated with this area of memory * * start: Area start address * end:   Area end address * flags: nonzero for I cache as well */		.align	5ENTRY(cpu_arm922_cache_clean_invalidate_range)	bic	r0, r0, #DCACHELINESIZE - 1	@ && added by PGM	bic	r1, r1, #DCACHELINESIZE - 1     @ && added by DHM	sub	r3, r1, r0	cmp	r3, #MAX_AREA_SIZE	bgt	cpu_arm922_cache_clean_invalidate_all_r21:	teq	r2, #0#ifdef CONFIG_CPU_ARM922_WRITETHROUGH	mcr	p15, 0, r0, c7, c6, 1		@ invalidate D entry	mcrne	p15, 0, r0, c7, c5, 1		@ invalidate I entry	add	r0, r0, #DCACHELINESIZE	mcr	p15, 0, r0, c7, c6, 1		@ invalidate D entry	mcrne	p15, 0, r0, c7, c5, 1		@ invalidate I entry	add	r0, r0, #DCACHELINESIZE#else	mcr	p15, 0, r0, c7, c14, 1		@ clean and invalidate D entry	mcrne	p15, 0, r0, c7, c5, 1		@ invalidate I entry	add	r0, r0, #DCACHELINESIZE	mcr	p15, 0, r0, c7, c14, 1		@ clean and invalidate D entry	mcrne	p15, 0, r0, c7, c5, 1		@ invalidate I entry	add	r0, r0, #DCACHELINESIZE#endif	cmp	r0, r1	blt	1b	mcr	p15, 0, r1, c7, c10, 4		@ drain WB	mov	pc, lr/* * cpu_arm922_flush_ram_page(page) * * clean and invalidate all cache lines associated with this area of memory * * page: page to clean and invalidate */	.align	5ENTRY(cpu_arm922_flush_ram_page)	mov	r1, #PAGESIZE#ifdef CONFIG_CPU_ARM922_WRITETHROUGH1:	mcr	p15, 0, r0, c7, c6, 1		@ invalidate D entry	add	r0, r0, #DCACHELINESIZE	mcr	p15, 0, r0, c7, c6, 1		@ invalidate D entry	add	r0, r0, #DCACHELINESIZE#else1:	mcr	p15, 0, r0, c7, c14, 1		@ clean and invalidate D entry	add	r0, r0, #DCACHELINESIZE	mcr	p15, 0, r0, c7, c14, 1		@ clean and invalidate D entry	add	r0, r0, #DCACHELINESIZE#endif	subs	r1, r1, #2 * DCACHELINESIZE	bne	1b	mcr	p15, 0, r1, c7, c10, 4		@ drain WB	mov	pc, lr/* ================================ D-CACHE =============================== *//* * cpu_arm922_dcache_invalidate_range(start, end) * * throw away all D-cached data in specified region without an obligation * to write them back.  Note however that we must clean the D-cached entries * around the boundaries if the start and/or end address are not cache * aligned. * * start: virtual start address * end:   virtual end address */	.align	5ENTRY(cpu_arm922_dcache_invalidate_range)#ifndef CONFIG_CPU_ARM922_WRITETHROUGH	tst	r0, #DCACHELINESIZE - 1	mcrne	p15, 0, r0, c7, c10, 1		@ clean D entry	tst	r1, #DCACHELINESIZE - 1	mcrne	p15, 0, r1, c7, c10, 1		@ clean D entry#endif		@ clean D entry	bic	r0, r0, #DCACHELINESIZE - 1	bic	r1, r1, #DCACHELINESIZE - 11:	mcr	p15, 0, r0, c7, c6, 1		@ invalidate D entry	add	r0, r0, #DCACHELINESIZE	cmp	r0, r1	blt	1b	mov	pc, lr/* * cpu_arm922_dcache_clean_range(start, end) * * For the specified virtual address range, ensure that all caches contain * clean data, such that peripheral accesses to the physical RAM fetch * correct data. * * start: virtual start address * end:   virtual end address */	.align	5ENTRY(cpu_arm922_dcache_clean_range)#ifndef CONFIG_CPU_ARM922_WRITETHROUGH	bic	r0, r0, #DCACHELINESIZE - 1	sub	r1, r1, r0	cmp	r1, #MAX_AREA_SIZE	mov	r2, #0	bgt	cpu_arm922_cache_clean_invalidate_all_r2	bic	r1, r1, #DCACHELINESIZE -1	add	r1, r1, #DCACHELINESIZE1:	mcr	p15, 0, r0, c7, c10, 1		@ clean D entry	add	r0, r0, #DCACHELINESIZE	subs	r1, r1, #DCACHELINESIZE	bpl	1b#endif	mcr	p15, 0, r2, c7, c10, 4		@ drain WB	mov	pc, lr/* * cpu_arm922_dcache_clean_page(page) * * Cleans a single page of dcache so that if we have any future aliased * mappings, they will be consistent at the time that they are created. * * page: virtual address of page to clean from dcache * * Note: *  1. we don't need to flush the write buffer in this case. *  2. we don't invalidate the entries since when we write the page *     out to disk, the entries may get reloaded into the cache. */	.align	5ENTRY(cpu_arm922_dcache_clean_page)#ifndef CONFIG_CPU_ARM922_WRITETHROUGH	mov	r1, #PAGESIZE1:	mcr	p15, 0, r0, c7, c10, 1		@ clean D entry	add	r0, r0, #DCACHELINESIZE	mcr	p15, 0, r0, c7, c10, 1		@ clean D entry	add	r0, r0, #DCACHELINESIZE	subs	r1, r1, #2 * DCACHELINESIZE	bne	1b#endif	mov	pc, lr/* * cpu_arm922_dcache_clean_entry(addr) * * Clean the specified entry of any caches such that the MMU * translation fetches will obtain correct data. * * addr: cache-unaligned virtual address */	.align	5

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -