⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 etrap.s

📁 优龙2410linux2.6.8内核源代码
💻 S
字号:
/* $Id: etrap.S,v 1.31 2000/01/08 16:38:18 anton Exp $ * etrap.S: Sparc trap window preparation for entry into the *          Linux kernel. * * Copyright (C) 1995 David S. Miller (davem@caip.rutgers.edu) */#include <asm/head.h>#include <asm/asi.h>#include <asm/contregs.h>#include <asm/page.h>#include <asm/psr.h>#include <asm/ptrace.h>#include <asm/winmacro.h>#include <asm/asmmacro.h>#include <asm/thread_info.h>/* Registers to not touch at all. */#define t_psr        l0 /* Set by caller */#define t_pc         l1 /* Set by caller */#define t_npc        l2 /* Set by caller */#define t_wim        l3 /* Set by caller */#define t_twinmask   l4 /* Set at beginning of this entry routine. */#define t_kstack     l5 /* Set right before pt_regs frame is built */#define t_retpc      l6 /* If you change this, change winmacro.h header file */#define t_systable   l7 /* Never touch this, could be the syscall table ptr. */#define curptr       g6 /* Set after pt_regs frame is built */	.text	.align 4	/* SEVEN WINDOW PATCH INSTRUCTIONS */	.globl	tsetup_7win_patch1, tsetup_7win_patch2	.globl	tsetup_7win_patch3, tsetup_7win_patch4	.globl	tsetup_7win_patch5, tsetup_7win_patch6tsetup_7win_patch1:	sll	%t_wim, 0x6, %t_wimtsetup_7win_patch2:	and	%g2, 0x7f, %g2tsetup_7win_patch3:	and	%g2, 0x7f, %g2tsetup_7win_patch4:	and	%g1, 0x7f, %g1tsetup_7win_patch5:	sll	%t_wim, 0x6, %t_wimtsetup_7win_patch6:	and	%g2, 0x7f, %g2	/* END OF PATCH INSTRUCTIONS */	/* At trap time, interrupts and all generic traps do the	 * following:	 *	 * rd	%psr, %l0	 * b	some_handler	 * rd	%wim, %l3	 * nop	 *	 * Then 'some_handler' if it needs a trap frame (ie. it has	 * to call c-code and the trap cannot be handled in-window)	 * then it does the SAVE_ALL macro in entry.S which does	 *	 * sethi	%hi(trap_setup), %l4	 * jmpl		%l4 + %lo(trap_setup), %l6	 * nop	 */	/* 2 3 4  window number	 * -----	 * O T S  mnemonic	 *	 * O == Current window before trap	 * T == Window entered when trap occurred	 * S == Window we will need to save if (1<<T) == %wim	 *	 * Before execution gets here, it must be guaranteed that	 * %l0 contains trap time %psr, %l1 and %l2 contain the	 * trap pc and npc, and %l3 contains the trap time %wim.	 */	.globl	trap_setup, tsetup_patch1, tsetup_patch2	.globl	tsetup_patch3, tsetup_patch4	.globl	tsetup_patch5, tsetup_patch6trap_setup:	/* Calculate mask of trap window.  See if from user	 * or kernel and branch conditionally.	 */	mov	1, %t_twinmask	andcc	%t_psr, PSR_PS, %g0		 ! fromsupv_p = (psr & PSR_PS)	be	trap_setup_from_user		 ! nope, from user mode	 sll	%t_twinmask, %t_psr, %t_twinmask ! t_twinmask = (1 << psr)	/* From kernel, allocate more kernel stack and	 * build a pt_regs trap frame.	 */	sub	%fp, (STACKFRAME_SZ + TRACEREG_SZ), %t_kstack	STORE_PT_ALL(t_kstack, t_psr, t_pc, t_npc, g2)	/* See if we are in the trap window. */	andcc	%t_twinmask, %t_wim, %g0	bne	trap_setup_kernel_spill		! in trap window, clean up	 nop	/* Trap from kernel with a window available.	 * Just do it...	 */	jmpl	%t_retpc + 0x8, %g0	! return to caller	 mov	%t_kstack, %sp		! jump onto new stacktrap_setup_kernel_spill:	ld	[%curptr + TI_UWINMASK], %g1	orcc	%g0, %g1, %g0	bne	trap_setup_user_spill	! there are some user windows, yuck	/* Spill from kernel, but only kernel windows, adjust	 * %wim and go.	 */	 srl	%t_wim, 0x1, %g2	! begin computation of new %wimtsetup_patch1:	sll	%t_wim, 0x7, %t_wim	! patched on 7 window Sparcs	or	%t_wim, %g2, %g2tsetup_patch2:	and	%g2, 0xff, %g2		! patched on 7 window Sparcs	save	%g0, %g0, %g0	/* Set new %wim value */	wr	%g2, 0x0, %wim	/* Save the kernel window onto the corresponding stack. */	STORE_WINDOW(sp)	restore	%g0, %g0, %g0	jmpl	%t_retpc + 0x8, %g0	! return to caller	 mov	%t_kstack, %sp		! and onto new kernel stack#define STACK_OFFSET (THREAD_SIZE - TRACEREG_SZ - STACKFRAME_SZ)trap_setup_from_user:	/* We can't use %curptr yet. */	LOAD_CURRENT(t_kstack, t_twinmask)	sethi	%hi(STACK_OFFSET), %t_twinmask	or	%t_twinmask, %lo(STACK_OFFSET), %t_twinmask	add	%t_kstack, %t_twinmask, %t_kstack	mov	1, %t_twinmask	sll	%t_twinmask, %t_psr, %t_twinmask ! t_twinmask = (1 << psr)	/* Build pt_regs frame. */	STORE_PT_ALL(t_kstack, t_psr, t_pc, t_npc, g2)#if 0	/* If we're sure every task_struct is THREAD_SIZE aligned,	   we can speed this up. */	sethi	%hi(STACK_OFFSET), %curptr	or	%curptr, %lo(STACK_OFFSET), %curptr	sub	%t_kstack, %curptr, %curptr#else	sethi	%hi(~(THREAD_SIZE - 1)), %curptr	and	%t_kstack, %curptr, %curptr#endif	/* Clear current_thread_info->w_saved */	st	%g0, [%curptr + TI_W_SAVED]	/* See if we are in the trap window. */	andcc	%t_twinmask, %t_wim, %g0	bne	trap_setup_user_spill		! yep we are	 orn	%g0, %t_twinmask, %g1		! negate trap win mask into %g1	/* Trap from user, but not into the invalid window.	 * Calculate new umask.  The way this works is,	 * any window from the %wim at trap time until	 * the window right before the one we are in now,	 * is a user window.  A diagram:	 *	 *      7 6 5 4 3 2 1 0    window number	 *      ---------------	 *        I     L T        mnemonic	 *	 * Window 'I' is the invalid window in our example,	 * window 'L' is the window the user was in when	 * the trap occurred, window T is the trap window	 * we are in now.  So therefore, windows 5, 4 and	 * 3 are user windows.  The following sequence	 * computes the user winmask to represent this.	 */	subcc	%t_wim, %t_twinmask, %g2	bneg,a	1f	 sub	%g2, 0x1, %g21:	andn	%g2, %t_twinmask, %g2tsetup_patch3:	and	%g2, 0xff, %g2			! patched on 7win Sparcs	st	%g2, [%curptr + TI_UWINMASK]	! store new umask	jmpl	%t_retpc + 0x8, %g0		! return to caller	 mov	%t_kstack, %sp			! and onto kernel stacktrap_setup_user_spill:	/* A spill occurred from either kernel or user mode	 * and there exist some user windows to deal with.	 * A mask of the currently valid user windows	 * is in %g1 upon entry to here.	 */tsetup_patch4:	and	%g1, 0xff, %g1		! patched on 7win Sparcs, mask	srl	%t_wim, 0x1, %g2	! compute new %wimtsetup_patch5:	sll	%t_wim, 0x7, %t_wim	! patched on 7win Sparcs	or	%t_wim, %g2, %g2	! %g2 is new %wimtsetup_patch6:	and	%g2, 0xff, %g2		! patched on 7win Sparcs	andn	%g1, %g2, %g1		! clear this bit in %g1	st	%g1, [%curptr + TI_UWINMASK]	save	%g0, %g0, %g0	wr	%g2, 0x0, %wim	/* Call MMU-architecture dependent stack checking	 * routine.	 */	.globl	tsetup_mmu_patchmetsetup_mmu_patchme:	b	tsetup_sun4c_stackchk	 andcc	%sp, 0x7, %g0	/* Architecture specific stack checking routines.  When either	 * of these routines are called, the globals are free to use	 * as they have been safely stashed on the new kernel stack	 * pointer.  Thus the definition below for simplicity.	 */#define glob_tmp     g1	.globl	tsetup_sun4c_stackchktsetup_sun4c_stackchk:	/* Done by caller: andcc %sp, 0x7, %g0 */	bne	trap_setup_user_stack_is_bolixed	 sra	%sp, 29, %glob_tmp	add	%glob_tmp, 0x1, %glob_tmp	andncc	%glob_tmp, 0x1, %g0	bne	trap_setup_user_stack_is_bolixed	 and	%sp, 0xfff, %glob_tmp		! delay slot	/* See if our dump area will be on more than one	 * page.	 */	add	%glob_tmp, 0x38, %glob_tmp	andncc	%glob_tmp, 0xff8, %g0	be	tsetup_sun4c_onepage		! only one page to check	 lda	[%sp] ASI_PTE, %glob_tmp	! have to check first page anywaystsetup_sun4c_twopages:	/* Is first page ok permission wise? */	srl	%glob_tmp, 29, %glob_tmp	cmp	%glob_tmp, 0x6	bne	trap_setup_user_stack_is_bolixed	 add	%sp, 0x38, %glob_tmp		/* Is second page in vma hole? */	sra	%glob_tmp, 29, %glob_tmp	add	%glob_tmp, 0x1, %glob_tmp	andncc	%glob_tmp, 0x1, %g0	bne	trap_setup_user_stack_is_bolixed	 add	%sp, 0x38, %glob_tmp	lda	[%glob_tmp] ASI_PTE, %glob_tmptsetup_sun4c_onepage:	srl	%glob_tmp, 29, %glob_tmp	cmp	%glob_tmp, 0x6				! can user write to it?	bne	trap_setup_user_stack_is_bolixed	! failure	 nop	STORE_WINDOW(sp)	restore %g0, %g0, %g0	jmpl	%t_retpc + 0x8, %g0	 mov	%t_kstack, %sp	.globl	tsetup_srmmu_stackchktsetup_srmmu_stackchk:	/* Check results of callers andcc %sp, 0x7, %g0 */	bne	trap_setup_user_stack_is_bolixed	 sethi   %hi(PAGE_OFFSET), %glob_tmp	cmp	%glob_tmp, %sp	bleu,a	1f	 lda	[%g0] ASI_M_MMUREGS, %glob_tmp		! read MMU controltrap_setup_user_stack_is_bolixed:	/* From user/kernel into invalid window w/bad user	 * stack. Save bad user stack, and return to caller.	 */	SAVE_BOLIXED_USER_STACK(curptr, g3)	restore	%g0, %g0, %g0	jmpl	%t_retpc + 0x8, %g0	 mov	%t_kstack, %sp1:	/* Clear the fault status and turn on the no_fault bit. */	or	%glob_tmp, 0x2, %glob_tmp		! or in no_fault bit	sta	%glob_tmp, [%g0] ASI_M_MMUREGS		! set it	/* Dump the registers and cross fingers. */	STORE_WINDOW(sp)	/* Clear the no_fault bit and check the status. */	andn	%glob_tmp, 0x2, %glob_tmp	sta	%glob_tmp, [%g0] ASI_M_MMUREGS	mov	AC_M_SFAR, %glob_tmp	lda	[%glob_tmp] ASI_M_MMUREGS, %g0	mov	AC_M_SFSR, %glob_tmp	lda	[%glob_tmp] ASI_M_MMUREGS, %glob_tmp	! save away status of winstore	andcc	%glob_tmp, 0x2, %g0			! did we fault?	bne	trap_setup_user_stack_is_bolixed	! failure	 nop	restore %g0, %g0, %g0	jmpl	%t_retpc + 0x8, %g0	 mov	%t_kstack, %sp

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -