⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 head.s

📁 上传linux-jx2410的源代码
💻 S
字号:
/* *  linux/arch/arm/boot/compressed/head.S * *  Copyright (C) 1996-2002 Russell King * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. */#include <linux/config.h>#include <linux/linkage.h>/* * Debugging stuff * * Note that these macros must not contain any code which is not * 100% relocatable.  Any attempt to do so will result in a crash. * Please select one of the following when turning on debugging. */#ifdef DEBUG#if defined(CONFIG_DEBUG_DC21285_PORT)		.macro	loadsp,	rb		mov	\rb, #0x7c000000		.endm		.macro	writeb,	rb		strb	\rb, [r3, #0x3f8]		.endm#elif defined(CONFIG_ARCH_RPC)		.macro	loadsp,	rb		mov	\rb, #0x03000000		orr	\rb, \rb, #0x00010000		.endm		.macro	writeb,	rb		strb	\rb, [r3, #0x3f8 << 2]		.endm#elif defined(CONFIG_ARCH_INTEGRATOR)		.macro	loadsp, rb		mov	\rb, #0x16000000		.endm		.macro	writeb, rb		strb	\rb, [r3, #0]		.endm#elif defined(CONFIG_ARCH_SA1100)		.macro	loadsp, rb		mov	\rb, #0x80000000	@ physical base address#  if defined(CONFIG_DEBUG_LL_SER3)		add	\rb, \rb, #0x00050000	@ Ser3#  else		add	\rb, \rb, #0x00010000	@ Ser1#  endif		.endm		.macro	writeb, rb/* * "The ARM peripheral bus does not support byte or half-word operations. * All reads and writes of the UART by the CPU should be wordwide." *   - SA-1100 Developer's Manual, August 1999 */		str	\rb, [r3, #0x14]	@ UTDR		.endm#else#error no serial architecture defined#endif#endif		.macro	kputc,val		mov	r0, \val		bl	putc		.endm		.macro	kphex,val,len		mov	r0, \val		mov	r1, #\len		bl	phex		.endm		.macro	debug_reloc_start#ifdef DEBUG		kputc	#'\n'		kphex	r6, 8		/* processor id */		kputc	#':'		kphex	r7, 8		/* architecture id */		kputc	#':'		mrc	p15, 0, r0, c1, c0		kphex	r0, 8		/* control reg */		kputc	#'\n'		kphex	r5, 8		/* decompressed kernel start */		kputc	#'-'		kphex	r8, 8		/* decompressed kernel end  */		kputc	#'>'		kphex	r4, 8		/* kernel execution address */		kputc	#'\n'#endif		.endm		.macro	debug_reloc_end#ifdef DEBUG		kphex	r5, 8		/* end of kernel */		kputc	#'\n'		mov	r0, r4		bl	memdump		/* dump 256 bytes at start of kernel */#endif		.endm		.section ".start", #alloc, #execinstr/* * sort out different calling conventions */		.alignstart:		.type	start,#function		.rept	8		mov	r0, r0		.endr		b	1f		.word	0x016f2818		@ Magic numbers to help the loader		.word	start			@ absolute load/run zImage address		.word	_edata			@ zImage end address1:		mov	r7, r1			@ save architecture ID		mov	r8, #0			@ save r0#ifndef __ARM_ARCH_2__		/*		 * Booting from Angel - need to enter SVC mode and disable		 * FIQs/IRQs (numeric definitions from angel arm.h source).		 * We only do this if we were in user mode on entry.		 */		mrs	r2, cpsr		@ get current mode		tst	r2, #3			@ not user?		bne	not_angel		mov	r0, #0x17		@ angel_SWIreason_EnterSVC		swi	0x123456		@ angel_SWI_ARMnot_angel:		mrs	r2, cpsr		@ turn off interrupts to		orr	r2, r2, #0xc0		@ prevent angel from running		msr	cpsr_c, r2#else		teqp	pc, #0x0c000003		@ turn off interrupts#endif		/*		 * Note that some cache flushing and other stuff may		 * be needed here - is there an Angel SWI call for this?		 */		/*		 * some architecture specific code can be inserted		 * by the linker here, but it should preserve r7 and r8.		 */		.text		adr	r0, LC0		ldmia	r0, {r1, r2, r3, r4, r5, r6, ip, sp}		subs	r0, r0, r1		@ calculate the delta offset		teq	r0, #0			@ if delta is zero, we're		beq	not_relocated		@ running at the address we						@ were linked at.		/*		 * We're running at a different address.  We need to fix		 * up various pointers:		 *   r5 - zImage base address		 *   r6 - GOT start		 *   ip - GOT end		 */		add	r5, r5, r0		add	r6, r6, r0		add	ip, ip, r0#ifndef CONFIG_ZBOOT_ROM		/*		 * If we're running fully PIC === CONFIG_ZBOOT_ROM = n,		 * we need to fix up pointers into the BSS region.		 *   r2 - BSS start		 *   r3 - BSS end		 *   sp - stack pointer		 */		add	r2, r2, r0		add	r3, r3, r0		add	sp, sp, r0		/*		 * Relocate all entries in the GOT table.		 */1:		ldr	r1, [r6, #0]		add	r1, r1, r0		str	r1, [r6], #4		cmp	r6, ip		blo	1b#else		/*		 * Relocate entries in the GOT table.  We only relocate		 * the entries that are outside the (relocated) BSS region.		 */1:		ldr	r1, [r6, #0]		cmp	r1, r2			@ entry < bss_start ||		cmphs	r3, r1			@ _end < entry		addlo	r1, r1, r0		str	r1, [r6], #4		cmp	r6, ip		blo	1b#endifnot_relocated:	mov	r0, #01:		str	r0, [r2], #4		@ clear bss		str	r0, [r2], #4		str	r0, [r2], #4		str	r0, [r2], #4		cmp	r2, r3		blo	1b		/*		 * The C runtime environment should now be setup		 * sufficiently.  Turn the cache on, set up some		 * pointers, and start decompressing.		 */		bl	cache_on		mov	r1, sp			@ malloc space above stack		add	r2, sp, #0x10000	@ 64k max/* * Check to see if we will overwrite ourselves. *   r4 = final kernel address *   r5 = start of this image *   r2 = end of malloc space (and therefore this image) * We basically want: *   r4 >= r2 -> OK *   r4 + image length <= r5 -> OK */		cmp	r4, r2		bhs	wont_overwrite		add	r0, r4, #4096*1024	@ 4MB largest kernel size		cmp	r0, r5		bls	wont_overwrite		mov	r5, r2			@ decompress after malloc space		mov	r0, r5		mov	r3, r7		bl	decompress_kernel		add	r0, r0, #127		bic	r0, r0, #127		@ align the kernel length/* * r0     = decompressed kernel length * r1-r3  = unused * r4     = kernel execution address * r5     = decompressed kernel start * r6     = processor ID * r7     = architecture ID * r8-r14 = unused */		add	r1, r5, r0		@ end of decompressed kernel		adr	r2, reloc_start		ldr	r3, LC1		add	r3, r2, r31:		ldmia	r2!, {r8 - r13}		@ copy relocation code		stmia	r1!, {r8 - r13}		ldmia	r2!, {r8 - r13}		stmia	r1!, {r8 - r13}		cmp	r2, r3		blo	1b		bl	cache_clean_flush		add	pc, r5, r0		@ call relocation code/* * We're not in danger of overwriting ourselves.  Do this the simple way. * * r4     = kernel execution address * r7     = architecture ID */wont_overwrite:	mov	r0, r4		mov	r3, r7		bl	decompress_kernel		b	call_kernel		.type	LC0, #objectLC0:		.word	LC0			@ r1		.word	__bss_start		@ r2		.word	_end			@ r3		.word	_load_addr		@ r4		.word	_start			@ r5		.word	_got_start		@ r6		.word	_got_end		@ ip		.word	user_stack+4096		@ spLC1:		.word	reloc_end - reloc_start		.size	LC0, . - LC0/* * Turn on the cache.  We need to setup some page tables so that we * can have both the I and D caches on. * * We place the page tables 16k down from the kernel execution address, * and we hope that nothing else is using it.  If we're using it, we * will go pop! * * On entry, *  r4 = kernel execution address *  r6 = processor ID *  r7 = architecture number *  r8 = run-time address of "start" * On exit, *  r1, r2, r3, r8, r9, r12 corrupted * This routine must preserve: *  r4, r5, r6, r7 */		.align	5cache_on:	mov	r3, #8			@ cache_on function		b	call_cache_fn__setup_mmu:		sub	r3, r4, #16384		@ Page directory size		bic	r3, r3, #0xff		@ Align the pointer		bic	r3, r3, #0x3f00/* * Initialise the page tables, turning on the cacheable and bufferable * bits for the RAM area only. */		mov	r0, r3		mov	r8, r0, lsr #18		mov	r8, r8, lsl #18		@ start of RAM		add	r9, r8, #0x10000000	@ a reasonable RAM size		mov	r1, #0x12		orr	r1, r1, #3 << 10		add	r2, r3, #163841:		cmp	r1, r8			@ if virt > start of RAM#ifdef CONFIG_XSCALE_CACHE_ERRATA		orrhs	r1, r1, #0x08           @ set cacheable, not bufferable#else		orrhs	r1, r1, #0x0c		@ set cacheable, bufferable#endif		cmp	r1, r9			@ if virt > end of RAM		bichs	r1, r1, #0x0c		@ clear cacheable, bufferable		str	r1, [r0], #4		@ 1:1 mapping		add	r1, r1, #1048576		teq	r0, r2		bne	1b/* * If ever we are running from Flash, then we surely want the cache * to be enabled also for our execution instance...  We map 2MB of it * so there is no map overlap problem for up to 1 MB compressed kernel. * If the execution is in RAM then we would only be duplicating the above. */		mov	r1, #0x1e		orr	r1, r1, #3 << 10		mov	r2, pc, lsr #20		orr	r1, r1, r2, lsl #20		add	r0, r3, r2, lsl #2		str	r1, [r0], #4		add	r1, r1, #1048576		str	r1, [r0]		mov	pc, lr__armv4_cache_on:		mov	r12, lr		bl	__setup_mmu		mov	r0, #0		mcr	p15, 0, r0, c7, c10, 4	@ drain write buffer		mcr	p15, 0, r0, c8, c7	@ flush I,D TLBs		mcr	p15, 0, r3, c2, c0	@ load page table pointer		mov	r0, #-1		mcr	p15, 0, r0, c3, c0	@ load domain access register		mrc	p15, 0, r0, c1, c0		orr	r0, r0, #0x1000		@ I-cache enable#ifndef DEBUG		orr	r0, r0, #0x003d		@ Write buffer, mmu#endif		mcr	p15, 0, r0, c1, c0		mov	pc, r12__arm6_cache_on:		mov	r12, lr		bl	__setup_mmu		mov	r0, #0		mcr	p15, 0, r0, c7, c0, 0	@ invalidate whole cache v3		mcr	p15, 0, r0, c5, c0, 0	@ invalidate whole TLB v3		mcr	p15, 0, r3, c2, c0	@ load page table pointer		mov	r0, #-1		mcr	p15, 0, r0, c3, c0		mov	r0, #0x3d		mcr	p15, 0, r0, c1, c0, 0		mov	pc, r12/* * All code following this line is relocatable.  It is relocated by * the above code to the end of the decompressed kernel image and * executed there.  During this time, we have no stacks. * * r0     = decompressed kernel length * r1-r3  = unused * r4     = kernel execution address * r5     = decompressed kernel start * r6     = processor ID * r7     = architecture ID * r8-r14 = unused */		.align	5reloc_start:	add	r8, r5, r0		debug_reloc_start		mov	r1, r41:		.rept	4		ldmia	r5!, {r0, r2, r3, r9 - r13}	@ relocate kernel		stmia	r1!, {r0, r2, r3, r9 - r13}		.endr		cmp	r5, r8		blo	1b		debug_reloc_endcall_kernel:	bl	cache_clean_flush		bl	cache_off		mov	r0, #0		mov	r1, r7			@ restore architecture number		mov	pc, r4			@ call kernel/* * Here follow the relocatable cache support functions for the * various processors.  This is a generic hook for locating an * entry and jumping to an instruction at the specified offset * from the start of the block.  Please note this is all position * independent code. * *  r1  = corrupted *  r2  = corrupted *  r3  = block offset *  r6  = corrupted *  r12 = corrupted */call_cache_fn:	adr	r12, proc_types		mrc	p15, 0, r6, c0, c0	@ get processor ID1:		ldr	r1, [r12, #0]		@ get value		ldr	r2, [r12, #4]		@ get mask		eor	r1, r1, r6		@ (real ^ match)		tst	r1, r2			@       & mask		addeq	pc, r12, r3		@ call cache function		add	r12, r12, #4*5		b	1b/* * Table for cache operations.  This is basically: *   - CPU ID match *   - CPU ID mask *   - 'cache on' method instruction *   - 'cache off' method instruction *   - 'cache flush' method instruction * * We match an entry using: ((real_id ^ match) & mask) == 0 * * Writethrough caches generally only need 'on' and 'off' * methods.  Writeback caches _must_ have the flush method * defined. */		.type	proc_types,#objectproc_types:		.word	0x41560600		@ ARM6/610		.word	0xffffffe0		b	__arm6_cache_off	@ works, but slow		b	__arm6_cache_off		mov	pc, lr@		b	__arm6_cache_on		@ untested@		b	__arm6_cache_off@		b	__armv3_cache_flush		.word	0x41007000		@ ARM7/710		.word	0xfff8fe00		b	__arm7_cache_off		b	__arm7_cache_off		mov	pc, lr		.word	0x41807200		@ ARM720T (writethrough)		.word	0xffffff00		b	__armv4_cache_on		b	__armv4_cache_off		mov	pc, lr		.word	0x41129200		@ ARM920T		.word	0xff00fff0		b	__armv4_cache_on		b	__armv4_cache_off		b	__armv4_cache_flush		.word	0x4401a100		@ sa110 / sa1100		.word	0xffffffe0		b	__armv4_cache_on		b	__armv4_cache_off		b	__armv4_cache_flush		.word	0x6901b110		@ sa1110		.word	0xfffffff0		b	__armv4_cache_on		b	__armv4_cache_off		b	__armv4_cache_flush		.word	0x69050000		@ xscale		.word	0xffff0000		b	__armv4_cache_on		b	__armv4_cache_off		b	__armv4_cache_flush		.word	0			@ unrecognised type		.word	0		mov	pc, lr		mov	pc, lr		mov	pc, lr		.size	proc_types, . - proc_types/* * Turn off the Cache and MMU.  ARMv3 does not support * reading the control register, but ARMv4 does. * * On entry,  r6 = processor ID * On exit,   r0, r1, r2, r3, r12 corrupted * This routine must preserve: r4, r6, r7 */		.align	5cache_off:	mov	r3, #12			@ cache_off function		b	call_cache_fn__armv4_cache_off:		mrc	p15, 0, r0, c1, c0		bic	r0, r0, #0x000d		mcr	p15, 0, r0, c1, c0	@ turn MMU and cache off		mov	r0, #0		mcr	p15, 0, r0, c7, c7	@ invalidate whole cache v4		mcr	p15, 0, r0, c8, c7	@ invalidate whole TLB v4		mov	pc, lr__arm6_cache_off:		mov	r0, #0x00000030		@ ARM6 control reg.		b	__armv3_cache_off__arm7_cache_off:		mov	r0, #0x00000070		@ ARM7 control reg.		b	__armv3_cache_off__armv3_cache_off:		mcr	p15, 0, r0, c1, c0, 0	@ turn MMU and cache off		mov	r0, #0		mcr	p15, 0, r0, c7, c0, 0	@ invalidate whole cache v3		mcr	p15, 0, r0, c5, c0, 0	@ invalidate whole TLB v3		mov	pc, lr/* * Clean and flush the cache to maintain consistency. * * On entry, *  r6 = processor ID * On exit, *  r1, r2, r3, r12 corrupted * This routine must preserve: *  r0, r4, r5, r6, r7 */		.align	5cache_clean_flush:		mov	r3, #16		b	call_cache_fn__armv4_cache_flush:		bic	r1, pc, #31		add	r2, r1, #65536		@ 2x the largest dcache size1:		ldr	r12, [r1], #32		@ s/w flush D cache		teq	r1, r2		bne	1b		mcr	p15, 0, r1, c7, c7, 0	@ flush I cache		mcr	p15, 0, r1, c7, c10, 4	@ drain WB		mov	pc, lr__armv3_cache_flush:		mov	r1, #0		mcr	p15, 0, r0, c7, c0, 0	@ invalidate whole cache v3		mov	pc, lr/* * Various debugging routines for printing hex characters and * memory, which again must be relocatable. */#ifdef DEBUG		.type	phexbuf,#objectphexbuf:	.space	12		.size	phexbuf, . - phexbufphex:		adr	r3, phexbuf		mov	r2, #0		strb	r2, [r3, r1]1:		subs	r1, r1, #1		movmi	r0, r3		bmi	puts		and	r2, r0, #15		mov	r0, r0, lsr #4		cmp	r2, #10		addge	r2, r2, #7		add	r2, r2, #'0'		strb	r2, [r3, r1]		b	1bputs:		loadsp	r31:		ldrb	r2, [r0], #1		teq	r2, #0		moveq	pc, lr2:		writeb	r2		mov	r1, #0x000200003:		subs	r1, r1, #1		bne	3b		teq	r2, #'\n'		moveq	r2, #'\r'		beq	2b		teq	r0, #0		bne	1b		mov	pc, lrputc:		mov	r2, r0		mov	r0, #0		loadsp	r3		b	2bmemdump:	mov	r12, r0		mov	r10, lr		mov	r11, #02:		mov	r0, r11, lsl #2		add	r0, r0, r12		mov	r1, #8		bl	phex		mov	r0, #':'		bl	putc1:		mov	r0, #' '		bl	putc		ldr	r0, [r12, r11, lsl #2]		mov	r1, #8		bl	phex		and	r0, r11, #7		teq	r0, #3		moveq	r0, #' '		bleq	putc		and	r0, r11, #7		add	r11, r11, #1		teq	r0, #7		bne	1b		mov	r0, #'\n'		bl	putc		cmp	r11, #64		blt	2b		mov	pc, r10#endifreloc_end:		.align		.section ".stack", "aw"user_stack:	.space	4096

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -