⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 tlb.c

📁 linux-2.6.15.6
💻 C
字号:
/* arch/sparc64/mm/tlb.c * * Copyright (C) 2004 David S. Miller <davem@redhat.com> */#include <linux/kernel.h>#include <linux/init.h>#include <linux/percpu.h>#include <linux/mm.h>#include <linux/swap.h>#include <asm/pgtable.h>#include <asm/pgalloc.h>#include <asm/tlbflush.h>#include <asm/cacheflush.h>#include <asm/mmu_context.h>#include <asm/tlb.h>/* Heavily inspired by the ppc64 code.  */DEFINE_PER_CPU(struct mmu_gather, mmu_gathers) = { 0, };void flush_tlb_pending(void){	struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);	if (mp->tlb_nr) {		if (CTX_VALID(mp->mm->context)) {#ifdef CONFIG_SMP			smp_flush_tlb_pending(mp->mm, mp->tlb_nr,					      &mp->vaddrs[0]);#else			__flush_tlb_pending(CTX_HWBITS(mp->mm->context),					    mp->tlb_nr, &mp->vaddrs[0]);#endif		}		mp->tlb_nr = 0;	}}void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, pte_t *ptep, pte_t orig){	struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);	unsigned long nr;	vaddr &= PAGE_MASK;	if (pte_exec(orig))		vaddr |= 0x1UL;	if (pte_dirty(orig)) {		unsigned long paddr, pfn = pte_pfn(orig);		struct address_space *mapping;		struct page *page;		if (!pfn_valid(pfn))			goto no_cache_flush;		page = pfn_to_page(pfn);		if (PageReserved(page))			goto no_cache_flush;		/* A real file page? */		mapping = page_mapping(page);		if (!mapping)			goto no_cache_flush;		paddr = (unsigned long) page_address(page);		if ((paddr ^ vaddr) & (1 << 13))			flush_dcache_page_all(mm, page);	}no_cache_flush:	if (mp->fullmm)		return;	nr = mp->tlb_nr;	if (unlikely(nr != 0 && mm != mp->mm)) {		flush_tlb_pending();		nr = 0;	}	if (nr == 0)		mp->mm = mm;	mp->vaddrs[nr] = vaddr;	mp->tlb_nr = ++nr;	if (nr >= TLB_BATCH_NR)		flush_tlb_pending();}void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end){	struct mmu_gather *mp = &__get_cpu_var(mmu_gathers);	unsigned long nr = mp->tlb_nr;	long s = start, e = end, vpte_base;	if (mp->fullmm)		return;	/* If start is greater than end, that is a real problem.  */	BUG_ON(start > end);	/* However, straddling the VA space hole is quite normal. */	s &= PMD_MASK;	e = (e + PMD_SIZE - 1) & PMD_MASK;	vpte_base = (tlb_type == spitfire ?		     VPTE_BASE_SPITFIRE :		     VPTE_BASE_CHEETAH);	if (unlikely(nr != 0 && mm != mp->mm)) {		flush_tlb_pending();		nr = 0;	}	if (nr == 0)		mp->mm = mm;	start = vpte_base + (s >> (PAGE_SHIFT - 3));	end = vpte_base + (e >> (PAGE_SHIFT - 3));	/* If the request straddles the VA space hole, we	 * need to swap start and end.  The reason this	 * occurs is that "vpte_base" is the center of	 * the linear page table mapping area.  Thus,	 * high addresses with the sign bit set map to	 * addresses below vpte_base and non-sign bit	 * addresses map to addresses above vpte_base.	 */	if (end < start) {		unsigned long tmp = start;		start = end;		end = tmp;	}	while (start < end) {		mp->vaddrs[nr] = start;		mp->tlb_nr = ++nr;		if (nr >= TLB_BATCH_NR) {			flush_tlb_pending();			nr = 0;		}		start += PAGE_SIZE;	}	if (nr)		flush_tlb_pending();}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -