pgtable.h
来自「linux 内核源代码」· C头文件 代码 · 共 964 行 · 第 1/2 页
H
964 行
/* * include/asm-s390/pgtable.h * * S390 version * Copyright (C) 1999,2000 IBM Deutschland Entwicklung GmbH, IBM Corporation * Author(s): Hartmut Penner (hp@de.ibm.com) * Ulrich Weigand (weigand@de.ibm.com) * Martin Schwidefsky (schwidefsky@de.ibm.com) * * Derived from "include/asm-i386/pgtable.h" */#ifndef _ASM_S390_PGTABLE_H#define _ASM_S390_PGTABLE_H/* * The Linux memory management assumes a three-level page table setup. For * s390 31 bit we "fold" the mid level into the top-level page table, so * that we physically have the same two-level page table as the s390 mmu * expects in 31 bit mode. For s390 64 bit we use three of the five levels * the hardware provides (region first and region second tables are not * used). * * The "pgd_xxx()" functions are trivial for a folded two-level * setup: the pgd is never bad, and a pmd always exists (as it's folded * into the pgd entry) * * This file contains the functions and defines necessary to modify and use * the S390 page table tree. */#ifndef __ASSEMBLY__#include <linux/mm_types.h>#include <asm/bug.h>#include <asm/processor.h>extern pgd_t swapper_pg_dir[] __attribute__ ((aligned (4096)));extern void paging_init(void);extern void vmem_map_init(void);/* * The S390 doesn't have any external MMU info: the kernel page * tables contain all the necessary information. */#define update_mmu_cache(vma, address, pte) do { } while (0)/* * ZERO_PAGE is a global shared page that is always zero: used * for zero-mapped memory areas etc.. */extern char empty_zero_page[PAGE_SIZE];#define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page))#endif /* !__ASSEMBLY__ *//* * PMD_SHIFT determines the size of the area a second-level page * table can map * PGDIR_SHIFT determines what a third-level page table entry can map */#ifndef __s390x__# define PMD_SHIFT 22# define PUD_SHIFT 22# define PGDIR_SHIFT 22#else /* __s390x__ */# define PMD_SHIFT 21# define PUD_SHIFT 31# define PGDIR_SHIFT 31#endif /* __s390x__ */#define PMD_SIZE (1UL << PMD_SHIFT)#define PMD_MASK (~(PMD_SIZE-1))#define PUD_SIZE (1UL << PUD_SHIFT)#define PUD_MASK (~(PUD_SIZE-1))#define PGDIR_SIZE (1UL << PGDIR_SHIFT)#define PGDIR_MASK (~(PGDIR_SIZE-1))/* * entries per page directory level: the S390 is two-level, so * we don't really have any PMD directory physically. * for S390 segment-table entries are combined to one PGD * that leads to 1024 pte per pgd */#ifndef __s390x__# define PTRS_PER_PTE 1024# define PTRS_PER_PMD 1# define PTRS_PER_PUD 1# define PTRS_PER_PGD 512#else /* __s390x__ */# define PTRS_PER_PTE 512# define PTRS_PER_PMD 1024# define PTRS_PER_PUD 1# define PTRS_PER_PGD 2048#endif /* __s390x__ */#define FIRST_USER_ADDRESS 0#define pte_ERROR(e) \ printk("%s:%d: bad pte %p.\n", __FILE__, __LINE__, (void *) pte_val(e))#define pmd_ERROR(e) \ printk("%s:%d: bad pmd %p.\n", __FILE__, __LINE__, (void *) pmd_val(e))#define pud_ERROR(e) \ printk("%s:%d: bad pud %p.\n", __FILE__, __LINE__, (void *) pud_val(e))#define pgd_ERROR(e) \ printk("%s:%d: bad pgd %p.\n", __FILE__, __LINE__, (void *) pgd_val(e))#ifndef __ASSEMBLY__/* * Just any arbitrary offset to the start of the vmalloc VM area: the * current 8MB value just means that there will be a 8MB "hole" after the * physical memory until the kernel virtual memory starts. That means that * any out-of-bounds memory accesses will hopefully be caught. * The vmalloc() routines leaves a hole of 4kB between each vmalloced * area for the same reason. ;) * vmalloc area starts at 4GB to prevent syscall table entry exchanging * from modules. */extern unsigned long vmalloc_end;#ifdef CONFIG_64BIT#define VMALLOC_ADDR (max(0x100000000UL, (unsigned long) high_memory))#else#define VMALLOC_ADDR ((unsigned long) high_memory)#endif#define VMALLOC_OFFSET (8*1024*1024)#define VMALLOC_START ((VMALLOC_ADDR + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1))#define VMALLOC_END vmalloc_end/* * We need some free virtual space to be able to do vmalloc. * VMALLOC_MIN_SIZE defines the minimum size of the vmalloc * area. On a machine with 2GB memory we make sure that we * have at least 128MB free space for vmalloc. On a machine * with 4TB we make sure we have at least 128GB. */#ifndef __s390x__#define VMALLOC_MIN_SIZE 0x8000000UL#define VMALLOC_END_INIT 0x80000000UL#else /* __s390x__ */#define VMALLOC_MIN_SIZE 0x2000000000UL#define VMALLOC_END_INIT 0x40000000000UL#endif /* __s390x__ *//* * A 31 bit pagetable entry of S390 has following format: * | PFRA | | OS | * 0 0IP0 * 00000000001111111111222222222233 * 01234567890123456789012345678901 * * I Page-Invalid Bit: Page is not available for address-translation * P Page-Protection Bit: Store access not possible for page * * A 31 bit segmenttable entry of S390 has following format: * | P-table origin | |PTL * 0 IC * 00000000001111111111222222222233 * 01234567890123456789012345678901 * * I Segment-Invalid Bit: Segment is not available for address-translation * C Common-Segment Bit: Segment is not private (PoP 3-30) * PTL Page-Table-Length: Page-table length (PTL+1*16 entries -> up to 256) * * The 31 bit segmenttable origin of S390 has following format: * * |S-table origin | | STL | * X **GPS * 00000000001111111111222222222233 * 01234567890123456789012345678901 * * X Space-Switch event: * G Segment-Invalid Bit: * * P Private-Space Bit: Segment is not private (PoP 3-30) * S Storage-Alteration: * STL Segment-Table-Length: Segment-table length (STL+1*16 entries -> up to 2048) * * A 64 bit pagetable entry of S390 has following format: * | PFRA |0IP0| OS | * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * I Page-Invalid Bit: Page is not available for address-translation * P Page-Protection Bit: Store access not possible for page * * A 64 bit segmenttable entry of S390 has following format: * | P-table origin | TT * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * I Segment-Invalid Bit: Segment is not available for address-translation * C Common-Segment Bit: Segment is not private (PoP 3-30) * P Page-Protection Bit: Store access not possible for page * TT Type 00 * * A 64 bit region table entry of S390 has following format: * | S-table origin | TF TTTL * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * I Segment-Invalid Bit: Segment is not available for address-translation * TT Type 01 * TF * TL Table length * * The 64 bit regiontable origin of S390 has following format: * | region table origon | DTTL * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * X Space-Switch event: * G Segment-Invalid Bit: * P Private-Space Bit: * S Storage-Alteration: * R Real space * TL Table-Length: * * A storage key has the following format: * | ACC |F|R|C|0| * 0 3 4 5 6 7 * ACC: access key * F : fetch protection bit * R : referenced bit * C : changed bit *//* Hardware bits in the page table entry */#define _PAGE_RO 0x200 /* HW read-only bit */#define _PAGE_INVALID 0x400 /* HW invalid bit *//* Software bits in the page table entry */#define _PAGE_SWT 0x001 /* SW pte type bit t */#define _PAGE_SWX 0x002 /* SW pte type bit x *//* Six different types of pages. */#define _PAGE_TYPE_EMPTY 0x400#define _PAGE_TYPE_NONE 0x401#define _PAGE_TYPE_SWAP 0x403#define _PAGE_TYPE_FILE 0x601 /* bit 0x002 is used for offset !! */#define _PAGE_TYPE_RO 0x200#define _PAGE_TYPE_RW 0x000#define _PAGE_TYPE_EX_RO 0x202#define _PAGE_TYPE_EX_RW 0x002/* * PTE type bits are rather complicated. handle_pte_fault uses pte_present, * pte_none and pte_file to find out the pte type WITHOUT holding the page * table lock. ptep_clear_flush on the other hand uses ptep_clear_flush to * invalidate a given pte. ipte sets the hw invalid bit and clears all tlbs * for the page. The page table entry is set to _PAGE_TYPE_EMPTY afterwards. * This change is done while holding the lock, but the intermediate step * of a previously valid pte with the hw invalid bit set can be observed by * handle_pte_fault. That makes it necessary that all valid pte types with * the hw invalid bit set must be distinguishable from the four pte types * empty, none, swap and file. * * irxt ipte irxt * _PAGE_TYPE_EMPTY 1000 -> 1000 * _PAGE_TYPE_NONE 1001 -> 1001 * _PAGE_TYPE_SWAP 1011 -> 1011 * _PAGE_TYPE_FILE 11?1 -> 11?1 * _PAGE_TYPE_RO 0100 -> 1100 * _PAGE_TYPE_RW 0000 -> 1000 * _PAGE_TYPE_EX_RO 0110 -> 1110 * _PAGE_TYPE_EX_RW 0010 -> 1010 * * pte_none is true for bits combinations 1000, 1010, 1100, 1110 * pte_present is true for bits combinations 0000, 0010, 0100, 0110, 1001 * pte_file is true for bits combinations 1101, 1111 * swap pte is 1011 and 0001, 0011, 0101, 0111 are invalid. */#ifndef __s390x__/* Bits in the segment table address-space-control-element */#define _ASCE_SPACE_SWITCH 0x80000000UL /* space switch event */#define _ASCE_ORIGIN_MASK 0x7ffff000UL /* segment table origin */#define _ASCE_PRIVATE_SPACE 0x100 /* private space control */#define _ASCE_ALT_EVENT 0x80 /* storage alteration event control */#define _ASCE_TABLE_LENGTH 0x7f /* 128 x 64 entries = 8k *//* Bits in the segment table entry */#define _SEGMENT_ENTRY_ORIGIN 0x7fffffc0UL /* page table origin */#define _SEGMENT_ENTRY_INV 0x20 /* invalid segment table entry */#define _SEGMENT_ENTRY_COMMON 0x10 /* common segment bit */#define _SEGMENT_ENTRY_PTL 0x0f /* page table length */#define _SEGMENT_ENTRY (_SEGMENT_ENTRY_PTL)#define _SEGMENT_ENTRY_EMPTY (_SEGMENT_ENTRY_INV)#else /* __s390x__ *//* Bits in the segment/region table address-space-control-element */#define _ASCE_ORIGIN ~0xfffUL/* segment table origin */#define _ASCE_PRIVATE_SPACE 0x100 /* private space control */#define _ASCE_ALT_EVENT 0x80 /* storage alteration event control */#define _ASCE_SPACE_SWITCH 0x40 /* space switch event */#define _ASCE_REAL_SPACE 0x20 /* real space control */#define _ASCE_TYPE_MASK 0x0c /* asce table type mask */#define _ASCE_TYPE_REGION1 0x0c /* region first table type */#define _ASCE_TYPE_REGION2 0x08 /* region second table type */#define _ASCE_TYPE_REGION3 0x04 /* region third table type */#define _ASCE_TYPE_SEGMENT 0x00 /* segment table type */#define _ASCE_TABLE_LENGTH 0x03 /* region table length *//* Bits in the region table entry */#define _REGION_ENTRY_ORIGIN ~0xfffUL/* region/segment table origin */#define _REGION_ENTRY_INV 0x20 /* invalid region table entry */#define _REGION_ENTRY_TYPE_MASK 0x0c /* region/segment table type mask */#define _REGION_ENTRY_TYPE_R1 0x0c /* region first table type */#define _REGION_ENTRY_TYPE_R2 0x08 /* region second table type */#define _REGION_ENTRY_TYPE_R3 0x04 /* region third table type */#define _REGION_ENTRY_LENGTH 0x03 /* region third length */#define _REGION1_ENTRY (_REGION_ENTRY_TYPE_R1 | _REGION_ENTRY_LENGTH)#define _REGION1_ENTRY_EMPTY (_REGION_ENTRY_TYPE_R1 | _REGION_ENTRY_INV)#define _REGION2_ENTRY (_REGION_ENTRY_TYPE_R2 | _REGION_ENTRY_LENGTH)#define _REGION2_ENTRY_EMPTY (_REGION_ENTRY_TYPE_R2 | _REGION_ENTRY_INV)#define _REGION3_ENTRY (_REGION_ENTRY_TYPE_R3 | _REGION_ENTRY_LENGTH)#define _REGION3_ENTRY_EMPTY (_REGION_ENTRY_TYPE_R3 | _REGION_ENTRY_INV)/* Bits in the segment table entry */#define _SEGMENT_ENTRY_ORIGIN ~0x7ffUL/* segment table origin */#define _SEGMENT_ENTRY_RO 0x200 /* page protection bit */#define _SEGMENT_ENTRY_INV 0x20 /* invalid segment table entry */#define _SEGMENT_ENTRY (0)#define _SEGMENT_ENTRY_EMPTY (_SEGMENT_ENTRY_INV)#endif /* __s390x__ *//* * A user page table pointer has the space-switch-event bit, the * private-space-control bit and the storage-alteration-event-control * bit set. A kernel page table pointer doesn't need them. */#define _ASCE_USER_BITS (_ASCE_SPACE_SWITCH | _ASCE_PRIVATE_SPACE | \ _ASCE_ALT_EVENT)/* Bits int the storage key */#define _PAGE_CHANGED 0x02 /* HW changed bit */#define _PAGE_REFERENCED 0x04 /* HW referenced bit *//* * Page protection definitions. */#define PAGE_NONE __pgprot(_PAGE_TYPE_NONE)#define PAGE_RO __pgprot(_PAGE_TYPE_RO)#define PAGE_RW __pgprot(_PAGE_TYPE_RW)#define PAGE_EX_RO __pgprot(_PAGE_TYPE_EX_RO)#define PAGE_EX_RW __pgprot(_PAGE_TYPE_EX_RW)#define PAGE_KERNEL PAGE_RW#define PAGE_COPY PAGE_RO/* * Dependent on the EXEC_PROTECT option s390 can do execute protection. * Write permission always implies read permission. In theory with a * primary/secondary page table execute only can be implemented but * it would cost an additional bit in the pte to distinguish all the * different pte types. To avoid that execute permission currently * implies read permission as well. */ /*xwr*/#define __P000 PAGE_NONE#define __P001 PAGE_RO#define __P010 PAGE_RO#define __P011 PAGE_RO#define __P100 PAGE_EX_RO#define __P101 PAGE_EX_RO#define __P110 PAGE_EX_RO#define __P111 PAGE_EX_RO#define __S000 PAGE_NONE#define __S001 PAGE_RO#define __S010 PAGE_RW#define __S011 PAGE_RW#define __S100 PAGE_EX_RO#define __S101 PAGE_EX_RO#define __S110 PAGE_EX_RW#define __S111 PAGE_EX_RW#ifndef __s390x__# define PxD_SHADOW_SHIFT 1#else /* __s390x__ */# define PxD_SHADOW_SHIFT 2#endif /* __s390x__ */static inline struct page *get_shadow_page(struct page *page){ if (s390_noexec && page->index) return virt_to_page((void *)(addr_t) page->index); return NULL;}static inline void *get_shadow_pte(void *table){ unsigned long addr, offset; struct page *page; addr = (unsigned long) table; offset = addr & (PAGE_SIZE - 1); page = virt_to_page((void *)(addr ^ offset)); return (void *)(addr_t)(page->index ? (page->index | offset) : 0UL);}static inline void *get_shadow_table(void *table){ unsigned long addr, offset; struct page *page; addr = (unsigned long) table; offset = addr & ((PAGE_SIZE << PxD_SHADOW_SHIFT) - 1); page = virt_to_page((void *)(addr ^ offset)); return (void *)(addr_t)(page->index ? (page->index | offset) : 0UL);}/* * Certain architectures need to do special things when PTEs * within a page table are directly modified. Thus, the following * hook is made available. */static inline void set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *pteptr, pte_t pteval){ pte_t *shadow_pte = get_shadow_pte(pteptr); *pteptr = pteval; if (shadow_pte) { if (!(pte_val(pteval) & _PAGE_INVALID) && (pte_val(pteval) & _PAGE_SWX)) pte_val(*shadow_pte) = pte_val(pteval) | _PAGE_RO; else pte_val(*shadow_pte) = _PAGE_TYPE_EMPTY; }}/* * pgd/pmd/pte query functions */#ifndef __s390x__static inline int pgd_present(pgd_t pgd) { return 1; }static inline int pgd_none(pgd_t pgd) { return 0; }static inline int pgd_bad(pgd_t pgd) { return 0; }static inline int pud_present(pud_t pud) { return 1; }static inline int pud_none(pud_t pud) { return 0; }static inline int pud_bad(pud_t pud) { return 0; }#else /* __s390x__ */static inline int pgd_present(pgd_t pgd) { return 1; }static inline int pgd_none(pgd_t pgd) { return 0; }static inline int pgd_bad(pgd_t pgd) { return 0; }static inline int pud_present(pud_t pud){ return (pud_val(pud) & _REGION_ENTRY_ORIGIN) != 0UL;}static inline int pud_none(pud_t pud){ return (pud_val(pud) & _REGION_ENTRY_INV) != 0UL;}static inline int pud_bad(pud_t pud){ unsigned long mask = ~_REGION_ENTRY_ORIGIN & ~_REGION_ENTRY_INV; return (pud_val(pud) & mask) != _REGION3_ENTRY;}#endif /* __s390x__ */static inline int pmd_present(pmd_t pmd){ return (pmd_val(pmd) & _SEGMENT_ENTRY_ORIGIN) != 0UL;}static inline int pmd_none(pmd_t pmd){ return (pmd_val(pmd) & _SEGMENT_ENTRY_INV) != 0UL;}static inline int pmd_bad(pmd_t pmd)
⌨️ 快捷键说明
复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?