⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 ultra.s

📁 优龙2410linux2.6.8内核源代码
💻 S
字号:
/* $Id: ultra.S,v 1.72 2002/02/09 19:49:31 davem Exp $ * ultra.S: Don't expand these all over the place... * * Copyright (C) 1997, 2000 David S. Miller (davem@redhat.com) */#include <linux/config.h>#include <asm/asi.h>#include <asm/pgtable.h>#include <asm/page.h>#include <asm/spitfire.h>#include <asm/mmu_context.h>#include <asm/pil.h>#include <asm/head.h>#include <asm/thread_info.h>	/* Basically, most of the Spitfire vs. Cheetah madness	 * has to do with the fact that Cheetah does not support	 * IMMU flushes out of the secondary context.  Someone needs	 * to throw a south lake birthday party for the folks	 * in Microelectronics who refused to fix this shit.	 */	/* This file is meant to be read efficiently by the CPU, not humans.	 * Staraj sie tego nikomu nie pierdolnac...	 */	.text	.align		32	.globl		__flush_tlb_mm__flush_tlb_mm: /* %o0=(ctx & TAG_CONTEXT_BITS), %o1=SECONDARY_CONTEXT */	ldxa		[%o1] ASI_DMMU, %g2	cmp		%g2, %o0	bne,pn		%icc, __spitfire_flush_tlb_mm_slow	 mov		0x50, %g3	stxa		%g0, [%g3] ASI_DMMU_DEMAP	stxa		%g0, [%g3] ASI_IMMU_DEMAP	retl	 flush		%g6	nop	nop	nop	nop	nop	nop	nop	nop	.align		32	.globl		__flush_tlb_pending__flush_tlb_pending:	/* %o0 = context, %o1 = nr, %o2 = vaddrs[] */	rdpr		%pstate, %g5	sllx		%o1, 3, %o1	andn		%g5, PSTATE_IE, %g2	wrpr		%g2, %pstate	mov		SECONDARY_CONTEXT, %o4	ldxa		[%o4] ASI_DMMU, %g2	stxa		%o0, [%o4] ASI_DMMU1:	sub		%o1, (1 << 3), %o1	ldx		[%o2 + %o1], %o3	andcc		%o3, 1, %g0	andn		%o3, 1, %o3	be,pn		%icc, 2f	 or		%o3, 0x10, %o3	stxa		%g0, [%o3] ASI_IMMU_DEMAP2:	stxa		%g0, [%o3] ASI_DMMU_DEMAP	membar		#Sync	brnz,pt		%o1, 1b	 nop	stxa		%g2, [%o4] ASI_DMMU	flush		%g6	retl	 wrpr		%g5, 0x0, %pstate	.align		32	.globl		__flush_tlb_kernel_range__flush_tlb_kernel_range:	/* %o0=start, %o1=end */	cmp		%o0, %o1	be,pn		%xcc, 2f	 sethi		%hi(PAGE_SIZE), %o4	sub		%o1, %o0, %o3	sub		%o3, %o4, %o3	or		%o0, 0x20, %o0		! Nucleus1:	stxa		%g0, [%o0 + %o3] ASI_DMMU_DEMAP	stxa		%g0, [%o0 + %o3] ASI_IMMU_DEMAP	membar		#Sync	brnz,pt		%o3, 1b	 sub		%o3, %o4, %o32:	retl	 flush		%g6__spitfire_flush_tlb_mm_slow:	rdpr		%pstate, %g1	wrpr		%g1, PSTATE_IE, %pstate	stxa		%o0, [%o1] ASI_DMMU	stxa		%g0, [%g3] ASI_DMMU_DEMAP	stxa		%g0, [%g3] ASI_IMMU_DEMAP	flush		%g6	stxa		%g2, [%o1] ASI_DMMU	flush		%g6	retl	 wrpr		%g1, 0, %pstate/* * The following code flushes one page_size worth. */#if (PAGE_SHIFT == 13)#define ITAG_MASK 0xfe#elif (PAGE_SHIFT == 16)#define ITAG_MASK 0x7fe#else#error unsupported PAGE_SIZE#endif	.align		32	.globl		__flush_icache_page__flush_icache_page:	/* %o0 = phys_page */	sethi		%hi(1 << 13), %o2	! IC_set bit	mov		1, %g1	srlx		%o0, 5, %o0	clr		%o1			! IC_addr	sllx		%g1, 36, %g1	ldda		[%o1] ASI_IC_TAG, %o4	sub		%g1, 1, %g2	or		%o0, %g1, %o0		! VALID+phys-addr comparitor	sllx		%g2, 1, %g2	andn		%g2, ITAG_MASK, %g2	! IC_tag mask	nop	nop	nop	nop	nop	nop1:	addx		%g0, %g0, %g0	ldda		[%o1 + %o2] ASI_IC_TAG, %g4	addx		%g0, %g0, %g0	and		%o5, %g2, %g3	cmp		%g3, %o0	add		%o1, 0x20, %o1	ldda		[%o1] ASI_IC_TAG, %o4	be,pn		%xcc, iflush12:	 nop	and		%g5, %g2, %g5	cmp		%g5, %o0	be,pn		%xcc, iflush23:	 cmp		%o1, %o2	bne,pt		%xcc, 1b	 addx		%g0, %g0, %g0	nop	retl	 ldx		[%g6 + TI_TASK], %g4iflush1:sub		%o1, 0x20, %g3	stxa		%g0, [%g3] ASI_IC_TAG	flush		%g6	ba,a,pt		%xcc, 2biflush2:sub		%o1, 0x20, %g3	stxa		%g0, [%o1 + %o2] ASI_IC_TAG	flush		%g6	ba,a,pt		%xcc, 3b#if (PAGE_SHIFT == 13)#define DTAG_MASK 0x3#elif (PAGE_SHIFT == 16)#define DTAG_MASK 0x1f#elif (PAGE_SHIFT == 19)#define DTAG_MASK 0xff#elif (PAGE_SHIFT == 22)#define DTAG_MASK 0x3ff#endif	.align		64	.globl		__flush_dcache_page__flush_dcache_page:	/* %o0=kaddr, %o1=flush_icache */	sethi		%uhi(PAGE_OFFSET), %g1	sllx		%g1, 32, %g1	sub		%o0, %g1, %o0	clr		%o4	srlx		%o0, 11, %o0	sethi		%hi(1 << 14), %o21:	ldxa		[%o4] ASI_DCACHE_TAG, %o3	! LSU	Group	add		%o4, (1 << 5), %o4		! IEU0	ldxa		[%o4] ASI_DCACHE_TAG, %g1	! LSU	Group	add		%o4, (1 << 5), %o4		! IEU0	ldxa		[%o4] ASI_DCACHE_TAG, %g2	! LSU	Group	o3 available	add		%o4, (1 << 5), %o4		! IEU0	andn		%o3, DTAG_MASK, %o3		! IEU1	ldxa		[%o4] ASI_DCACHE_TAG, %g3	! LSU	Group	add		%o4, (1 << 5), %o4		! IEU0	andn		%g1, DTAG_MASK, %g1		! IEU1	cmp		%o0, %o3			! IEU1	Group	be,a,pn		%xcc, dflush1			! CTI	 sub		%o4, (4 << 5), %o4		! IEU0	(Group)	cmp		%o0, %g1			! IEU1	Group	andn		%g2, DTAG_MASK, %g2		! IEU0	be,a,pn		%xcc, dflush2			! CTI	 sub		%o4, (3 << 5), %o4		! IEU0	(Group)	cmp		%o0, %g2			! IEU1	Group	andn		%g3, DTAG_MASK, %g3		! IEU0	be,a,pn		%xcc, dflush3			! CTI	 sub		%o4, (2 << 5), %o4		! IEU0	(Group)	cmp		%o0, %g3			! IEU1	Group	be,a,pn		%xcc, dflush4			! CTI	 sub		%o4, (1 << 5), %o4		! IEU02:	cmp		%o4, %o2			! IEU1	Group	bne,pt		%xcc, 1b			! CTI	 nop						! IEU0	/* The I-cache does not snoop local stores so we	 * better flush that too when necessary.	 */	brnz,pt		%o1, __flush_icache_page	 sllx		%o0, 11, %o0	retl	 nopdflush1:stxa		%g0, [%o4] ASI_DCACHE_TAG	add		%o4, (1 << 5), %o4dflush2:stxa		%g0, [%o4] ASI_DCACHE_TAG	add		%o4, (1 << 5), %o4dflush3:stxa		%g0, [%o4] ASI_DCACHE_TAG	add		%o4, (1 << 5), %o4dflush4:stxa		%g0, [%o4] ASI_DCACHE_TAG	add		%o4, (1 << 5), %o4	membar		#Sync	ba,pt		%xcc, 2b	 nop	.align		32__prefill_dtlb:	rdpr		%pstate, %g7	wrpr		%g7, PSTATE_IE, %pstate	mov		TLB_TAG_ACCESS, %g1	stxa		%o5, [%g1] ASI_DMMU	stxa		%o2, [%g0] ASI_DTLB_DATA_IN	flush		%g6	retl	 wrpr		%g7, %pstate__prefill_itlb:	rdpr		%pstate, %g7	wrpr		%g7, PSTATE_IE, %pstate	mov		TLB_TAG_ACCESS, %g1	stxa		%o5, [%g1] ASI_IMMU	stxa		%o2, [%g0] ASI_ITLB_DATA_IN	flush		%g6	retl	 wrpr		%g7, %pstate	.globl		__update_mmu_cache__update_mmu_cache:	/* %o0=hw_context, %o1=address, %o2=pte, %o3=fault_code */	srlx		%o1, PAGE_SHIFT, %o1	andcc		%o3, FAULT_CODE_DTLB, %g0	sllx		%o1, PAGE_SHIFT, %o5	bne,pt		%xcc, __prefill_dtlb	 or		%o5, %o0, %o5	ba,a,pt		%xcc, __prefill_itlb	/* Cheetah specific versions, patched at boot time.  */__cheetah_flush_tlb_mm: /* 15 insns */	rdpr		%pstate, %g5	andn		%g5, PSTATE_IE, %g2	wrpr		%g2, 0x0, %pstate	wrpr		%g0, 1, %tl	mov		PRIMARY_CONTEXT, %o2	mov		0x40, %g3	ldxa		[%o2] ASI_DMMU, %g2	stxa		%o0, [%o2] ASI_DMMU	stxa		%g0, [%g3] ASI_DMMU_DEMAP	stxa		%g0, [%g3] ASI_IMMU_DEMAP	stxa		%g2, [%o2] ASI_DMMU	flush		%g6	wrpr		%g0, 0, %tl	retl	 wrpr		%g5, 0x0, %pstate__cheetah_flush_tlb_pending:	/* 22 insns */	/* %o0 = context, %o1 = nr, %o2 = vaddrs[] */	rdpr		%pstate, %g5	sllx		%o1, 3, %o1	andn		%g5, PSTATE_IE, %g2	wrpr		%g2, 0x0, %pstate	wrpr		%g0, 1, %tl	mov		PRIMARY_CONTEXT, %o4	ldxa		[%o4] ASI_DMMU, %g2	stxa		%o0, [%o4] ASI_DMMU1:	sub		%o1, (1 << 3), %o1	ldx		[%o2 + %o1], %o3	andcc		%o3, 1, %g0	be,pn		%icc, 2f	 andn		%o3, 1, %o3	stxa		%g0, [%o3] ASI_IMMU_DEMAP2:	stxa		%g0, [%o3] ASI_DMMU_DEMAP		brnz,pt		%o1, 1b	 membar		#Sync	stxa		%g2, [%o4] ASI_DMMU	flush		%g6	wrpr		%g0, 0, %tl	retl	 wrpr		%g5, 0x0, %pstateflush_dcpage_cheetah: /* 11 insns */	sethi		%uhi(PAGE_OFFSET), %g1	sllx		%g1, 32, %g1	sub		%o0, %g1, %o0	sethi		%hi(PAGE_SIZE), %o41:	subcc		%o4, (1 << 5), %o4	stxa		%g0, [%o0 + %o4] ASI_DCACHE_INVALIDATE	membar		#Sync	bne,pt		%icc, 1b	 nop	retl		/* I-cache flush never needed on Cheetah, see callers. */	 nopcheetah_patch_one:1:	lduw		[%o1], %g1	stw		%g1, [%o0]	flush		%o0	subcc		%o2, 1, %o2	add		%o1, 4, %o1	bne,pt		%icc, 1b	 add		%o0, 4, %o0	retl	 nop	.globl		cheetah_patch_cachetlbopscheetah_patch_cachetlbops:	save		%sp, -128, %sp	sethi		%hi(__flush_tlb_mm), %o0	or		%o0, %lo(__flush_tlb_mm), %o0	sethi		%hi(__cheetah_flush_tlb_mm), %o1	or		%o1, %lo(__cheetah_flush_tlb_mm), %o1	call		cheetah_patch_one	 mov		15, %o2	sethi		%hi(__flush_tlb_pending), %o0	or		%o0, %lo(__flush_tlb_pending), %o0	sethi		%hi(__cheetah_flush_tlb_pending), %o1	or		%o1, %lo(__cheetah_flush_tlb_pending), %o1	call		cheetah_patch_one	 mov		22, %o2	sethi		%hi(__flush_dcache_page), %o0	or		%o0, %lo(__flush_dcache_page), %o0	sethi		%hi(flush_dcpage_cheetah), %o1	or		%o1, %lo(flush_dcpage_cheetah), %o1	call		cheetah_patch_one	 mov		11, %o2	ret	 restore#ifdef CONFIG_SMP	/* These are all called by the slaves of a cross call, at	 * trap level 1, with interrupts fully disabled.	 *	 * Register usage:	 *   %g5	mm->context	(all tlb flushes)	 *   %g1	address arg 1	(tlb page and range flushes)	 *   %g7	address arg 2	(tlb range flush only)	 *	 *   %g6	ivector table, don't touch	 *   %g2	scratch 1	 *   %g3	scratch 2	 *   %g4	scratch 3	 *	 * TODO: Make xcall TLB range flushes use the tricks above... -DaveM	 */	.align		32	.globl		xcall_flush_tlb_mmxcall_flush_tlb_mm:	mov		PRIMARY_CONTEXT, %g2	mov		0x40, %g4	ldxa		[%g2] ASI_DMMU, %g3	stxa		%g5, [%g2] ASI_DMMU	stxa		%g0, [%g4] ASI_DMMU_DEMAP	stxa		%g0, [%g4] ASI_IMMU_DEMAP	stxa		%g3, [%g2] ASI_DMMU	retry	.globl		xcall_flush_tlb_pendingxcall_flush_tlb_pending:	/* %g5=context, %g1=nr, %g7=vaddrs[] */	sllx		%g1, 3, %g1	mov		PRIMARY_CONTEXT, %g4	ldxa		[%g4] ASI_DMMU, %g2	stxa		%g5, [%g4] ASI_DMMU1:	sub		%g1, (1 << 3), %g1	ldx		[%g7 + %g1], %g5	andcc		%g5, 0x1, %g0	be,pn		%icc, 2f	 andn		%g5, 0x1, %g5	stxa		%g0, [%g5] ASI_IMMU_DEMAP2:	stxa		%g0, [%g5] ASI_DMMU_DEMAP	membar		#Sync	brnz,pt		%g1, 1b	 nop	stxa		%g2, [%g4] ASI_DMMU	retry	.globl		xcall_flush_tlb_kernel_rangexcall_flush_tlb_kernel_range:	sethi		%hi(PAGE_SIZE - 1), %g2	or		%g2, %lo(PAGE_SIZE - 1), %g2	andn		%g1, %g2, %g1	andn		%g7, %g2, %g7	sub		%g7, %g1, %g3	add		%g2, 1, %g2	sub		%g3, %g2, %g3	or		%g1, 0x20, %g1		! Nucleus1:	stxa		%g0, [%g1 + %g3] ASI_DMMU_DEMAP	stxa		%g0, [%g1 + %g3] ASI_IMMU_DEMAP	membar		#Sync	brnz,pt		%g3, 1b	 sub		%g3, %g2, %g3	retry	nop	nop	/* This runs in a very controlled environment, so we do	 * not need to worry about BH races etc.	 */	.globl		xcall_sync_tickxcall_sync_tick:	rdpr		%pstate, %g2	wrpr		%g2, PSTATE_IG | PSTATE_AG, %pstate	rdpr		%pil, %g2	wrpr		%g0, 15, %pil	sethi		%hi(109f), %g7	b,pt		%xcc, etrap_irq109:	 or		%g7, %lo(109b), %g7	call		smp_synchronize_tick_client	 nop	clr		%l6	b		rtrap_xcall	 ldx		[%sp + PTREGS_OFF + PT_V9_TSTATE], %l1	/* NOTE: This is SPECIAL!!  We do etrap/rtrap however	 *       we choose to deal with the "BH's run with	 *       %pil==15" problem (described in asm/pil.h)	 *       by just invoking rtrap directly past where	 *       BH's are checked for.	 *	 *       We do it like this because we do not want %pil==15	 *       lockups to prevent regs being reported.	 */	.globl		xcall_report_regsxcall_report_regs:	rdpr		%pstate, %g2	wrpr		%g2, PSTATE_IG | PSTATE_AG, %pstate	rdpr		%pil, %g2	wrpr		%g0, 15, %pil	sethi		%hi(109f), %g7	b,pt		%xcc, etrap_irq109:	 or		%g7, %lo(109b), %g7	call		__show_regs	 add		%sp, PTREGS_OFF, %o0	clr		%l6	/* Has to be a non-v9 branch due to the large distance. */	b		rtrap_xcall	 ldx		[%sp + PTREGS_OFF + PT_V9_TSTATE], %l1	.align		32	.globl		xcall_flush_dcache_page_cheetahxcall_flush_dcache_page_cheetah: /* %g1 == physical page address */	sethi		%hi(PAGE_SIZE), %g31:	subcc		%g3, (1 << 5), %g3	stxa		%g0, [%g1 + %g3] ASI_DCACHE_INVALIDATE	membar		#Sync	bne,pt		%icc, 1b	 nop	retry	nop	.globl		xcall_flush_dcache_page_spitfirexcall_flush_dcache_page_spitfire: /* %g1 == physical page address				     %g7 == kernel page virtual address				     %g5 == (page->mapping != NULL)  */#if (L1DCACHE_SIZE > PAGE_SIZE)	srlx		%g1, (13 - 2), %g1	! Form tag comparitor	sethi		%hi(L1DCACHE_SIZE), %g3	! D$ size == 16K	sub		%g3, (1 << 5), %g3	! D$ linesize == 321:	ldxa		[%g3] ASI_DCACHE_TAG, %g2	andcc		%g2, 0x3, %g0	be,pn		%xcc, 2f	 andn		%g2, 0x3, %g2	cmp		%g2, %g1	bne,pt		%xcc, 2f	 nop	stxa		%g0, [%g3] ASI_DCACHE_TAG	membar		#Sync2:	cmp		%g3, 0	bne,pt		%xcc, 1b	 sub		%g3, (1 << 5), %g3	brz,pn		%g5, 2f#endif /* L1DCACHE_SIZE > PAGE_SIZE */	 sethi		%hi(PAGE_SIZE), %g31:	flush		%g7	subcc		%g3, (1 << 5), %g3	bne,pt		%icc, 1b	 add		%g7, (1 << 5), %g72:	retry	nop	nop	.globl		xcall_promstopxcall_promstop:	rdpr		%pstate, %g2	wrpr		%g2, PSTATE_IG | PSTATE_AG, %pstate	rdpr		%pil, %g2	wrpr		%g0, 15, %pil	sethi		%hi(109f), %g7	b,pt		%xcc, etrap_irq109:	 or		%g7, %lo(109b), %g7	flushw	call		prom_stopself	 nop	/* We should not return, just spin if we do... */1:	b,a,pt		%xcc, 1b	nop	.dataerrata32_hwbug:	.xword	0	.text	/* These two are not performance critical... */	.globl		xcall_flush_tlb_all_spitfirexcall_flush_tlb_all_spitfire:	/* Spitfire Errata #32 workaround. */	sethi		%hi(errata32_hwbug), %g4	stx		%g0, [%g4 + %lo(errata32_hwbug)]	clr		%g2	clr		%g31:	ldxa		[%g3] ASI_DTLB_DATA_ACCESS, %g4	and		%g4, _PAGE_L, %g5	brnz,pn		%g5, 2f	 mov		TLB_TAG_ACCESS, %g7	stxa		%g0, [%g7] ASI_DMMU	membar		#Sync	stxa		%g0, [%g3] ASI_DTLB_DATA_ACCESS	membar		#Sync	/* Spitfire Errata #32 workaround. */	sethi		%hi(errata32_hwbug), %g4	stx		%g0, [%g4 + %lo(errata32_hwbug)]2:	ldxa		[%g3] ASI_ITLB_DATA_ACCESS, %g4	and		%g4, _PAGE_L, %g5	brnz,pn		%g5, 2f	 mov		TLB_TAG_ACCESS, %g7	stxa		%g0, [%g7] ASI_IMMU	membar		#Sync	stxa		%g0, [%g3] ASI_ITLB_DATA_ACCESS	membar		#Sync	/* Spitfire Errata #32 workaround. */	sethi		%hi(errata32_hwbug), %g4	stx		%g0, [%g4 + %lo(errata32_hwbug)]2:	add		%g2, 1, %g2	cmp		%g2, SPITFIRE_HIGHEST_LOCKED_TLBENT	ble,pt		%icc, 1b	 sll		%g2, 3, %g3	flush		%g6	retry	.globl		xcall_flush_tlb_all_cheetahxcall_flush_tlb_all_cheetah:	mov		0x80, %g2	stxa		%g0, [%g2] ASI_DMMU_DEMAP	stxa		%g0, [%g2] ASI_IMMU_DEMAP	retry	/* These just get rescheduled to PIL vectors. */	.globl		xcall_call_functionxcall_call_function:	wr		%g0, (1 << PIL_SMP_CALL_FUNC), %set_softint	retry	.globl		xcall_receive_signalxcall_receive_signal:	wr		%g0, (1 << PIL_SMP_RECEIVE_SIGNAL), %set_softint	retry	.globl		xcall_capturexcall_capture:	wr		%g0, (1 << PIL_SMP_CAPTURE), %set_softint	retry#endif /* CONFIG_SMP */

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -