⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 pgtable.h

📁 this SRC packet is the headfiles that MIZI vivi bootloader needed when compling
💻 H
📖 第 1 页 / 共 2 页
字号:
/* *  include/asm-s390/pgtable.h * *  S390 64bit version *    Copyright (C) 2000 IBM Deutschland Entwicklung GmbH, IBM Corporation *    Author(s): Hartmut Penner (hp@de.ibm.com) *               Ulrich Weigand (weigand@de.ibm.com) *               Martin Schwidefsky (schwidefsky@de.ibm.com) * *  Derived from "include/asm-i386/pgtable.h" */#ifndef _ASM_S390_PGTABLE_H#define _ASM_S390_PGTABLE_H/* * The Linux memory management assumes a three-level page table setup. On * the S390, we use that, but "fold" the mid level into the top-level page * table, so that we physically have the same two-level page table as the * S390 mmu expects. * * This file contains the functions and defines necessary to modify and use * the S390 page table tree. */#ifndef __ASSEMBLY__#include <asm/processor.h>#include <linux/threads.h>extern pgd_t swapper_pg_dir[] __attribute__ ((aligned (4096)));extern void paging_init(void);/* Caches aren't brain-dead on S390. */#define flush_cache_all()                       do { } while (0)#define flush_cache_mm(mm)                      do { } while (0)#define flush_cache_range(mm, start, end)       do { } while (0)#define flush_cache_page(vma, vmaddr)           do { } while (0)#define flush_page_to_ram(page)                 do { } while (0)#define flush_dcache_page(page)			do { } while (0)#define flush_icache_range(start, end)          do { } while (0)#define flush_icache_page(vma,pg)               do { } while (0)/* * The S390 doesn't have any external MMU info: the kernel page * tables contain all the necessary information. */#define update_mmu_cache(vma, address, pte)     do { } while (0)/* * ZERO_PAGE is a global shared page that is always zero: used * for zero-mapped memory areas etc.. */extern char empty_zero_page[PAGE_SIZE];#define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page))#endif /* !__ASSEMBLY__ *//* * PMD_SHIFT determines the size of the area a second-level page * table can map */#define PMD_SHIFT       21#define PMD_SIZE        (1UL << PMD_SHIFT)#define PMD_MASK        (~(PMD_SIZE-1))/* PGDIR_SHIFT determines what a third-level page table entry can map */#define PGDIR_SHIFT     31#define PGDIR_SIZE      (1UL << PGDIR_SHIFT)#define PGDIR_MASK      (~(PGDIR_SIZE-1))/* * entries per page directory level: the S390 is two to five-level, * currently we use a 3 level lookup */#define PTRS_PER_PTE    512#define PTRS_PER_PMD    1024#define PTRS_PER_PGD    2048/* * pgd entries used up by user/kernel: */#define USER_PTRS_PER_PGD  2048#define USER_PGD_PTRS      2048#define KERNEL_PGD_PTRS    2048#define FIRST_USER_PGD_NR  0#define pte_ERROR(e) \	printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e))#define pmd_ERROR(e) \	printk("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))#define pgd_ERROR(e) \	printk("%s:%d: bad pgd %016lx.\n", __FILE__, __LINE__, pgd_val(e))#ifndef __ASSEMBLY__/* * Just any arbitrary offset to the start of the vmalloc VM area: the * current 8MB value just means that there will be a 8MB "hole" after the * physical memory until the kernel virtual memory starts.  That means that * any out-of-bounds memory accesses will hopefully be caught. * The vmalloc() routines leaves a hole of 4kB between each vmalloced * area for the same reason. ;) */#define VMALLOC_OFFSET  (8*1024*1024)#define VMALLOC_START   (((unsigned long) high_memory + VMALLOC_OFFSET) \                         & ~(VMALLOC_OFFSET-1))#define VMALLOC_VMADDR(x) ((unsigned long)(x))#define VMALLOC_END     (0x40000000000L)/* * A pagetable entry of S390 has following format: * |                     PFRA                         |0IP0|  OS  | * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * I Page-Invalid Bit:    Page is not available for address-translation * P Page-Protection Bit: Store access not possible for page * * A segmenttable entry of S390 has following format: * |        P-table origin                              |      TT * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * I Segment-Invalid Bit:    Segment is not available for address-translation * C Common-Segment Bit:     Segment is not private (PoP 3-30) * P Page-Protection Bit: Store access not possible for page * TT Type 00 * * A region table entry of S390 has following format: * |        S-table origin                             |   TF  TTTL * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * I Segment-Invalid Bit:    Segment is not available for address-translation * TT Type 01 * TF * TL Table lenght * * The regiontable origin of S390 has following format: * |      region table origon                          |       DTTL * 0000000000111111111122222222223333333333444444444455555555556666 * 0123456789012345678901234567890123456789012345678901234567890123 * * X Space-Switch event: * G Segment-Invalid Bit:   * P Private-Space Bit:     * S Storage-Alteration: * R Real space * TL Table-Length: * * A storage key has the following format: * | ACC |F|R|C|0| *  0   3 4 5 6 7 * ACC: access key * F  : fetch protection bit * R  : referenced bit * C  : changed bit *//* Bits in the page table entry */#define _PAGE_PRESENT   0x001          /* Software                         */#define _PAGE_MKCLEAR   0x002          /* Software                         */#define _PAGE_RO        0x200          /* HW read-only                     */#define _PAGE_INVALID   0x400          /* HW invalid                       *//* Bits in the segment table entry */#define _PMD_ENTRY_INV   0x20          /* invalid segment table entry      */#define _PMD_ENTRY       0x00        /* Bits in the region third table entry */#define _PGD_ENTRY_INV   0x20          /* invalid region table entry       */#define _PGD_ENTRY       0x07/* * User and kernel page directory */#define _REGION_THIRD       0x4#define _REGION_THIRD_LEN   0x3 #define _REGION_TABLE       (_REGION_THIRD|_REGION_THIRD_LEN|0x40|0x100)#define _KERN_REGION_TABLE  (_REGION_THIRD|_REGION_THIRD_LEN)/* Bits in the storage key */#define _PAGE_CHANGED    0x02          /* HW changed bit                   */#define _PAGE_REFERENCED 0x04          /* HW referenced bit                *//* * No mapping available */#define PAGE_INVALID    __pgprot(_PAGE_INVALID)#define PAGE_NONE       __pgprot(_PAGE_PRESENT | _PAGE_INVALID)#define PAGE_COPY       __pgprot(_PAGE_PRESENT | _PAGE_RO)#define PAGE_READONLY   __pgprot(_PAGE_PRESENT | _PAGE_RO)#define PAGE_SHARED     __pgprot(_PAGE_PRESENT )#define PAGE_KERNEL     __pgprot(_PAGE_PRESENT )/* * The S390 can't do page protection for execute, and considers that the * same are read. Also, write permissions imply read permissions. This is * the closest we can get.. */#define __P000  PAGE_NONE#define __P001  PAGE_READONLY#define __P010  PAGE_COPY#define __P011  PAGE_COPY#define __P100  PAGE_READONLY#define __P101  PAGE_READONLY#define __P110  PAGE_COPY#define __P111  PAGE_COPY#define __S000  PAGE_NONE#define __S001  PAGE_READONLY#define __S010  PAGE_SHARED#define __S011  PAGE_SHARED#define __S100  PAGE_READONLY#define __S101  PAGE_READONLY#define __S110  PAGE_SHARED#define __S111  PAGE_SHARED/* * Certain architectures need to do special things when PTEs * within a page table are directly modified.  Thus, the following * hook is made available. */extern inline void set_pte(pte_t *pteptr, pte_t pteval){	if ((pte_val(pteval) & (_PAGE_MKCLEAR|_PAGE_INVALID))	    == _PAGE_MKCLEAR) 	{		pte_val(pteval) &= ~_PAGE_MKCLEAR;               		asm volatile ("sske %0,%1" 				: : "d" (0), "a" (pte_val(pteval)));	}	*pteptr = pteval;}/* * Permanent address of a page. */#define page_address(page) ((page)->virtual)#define pages_to_mb(x) ((x) >> (20-PAGE_SHIFT))/* * pgd/pmd/pte query functions */extern inline int pgd_present(pgd_t pgd){	return (pgd_val(pgd) & ~PAGE_MASK) == _PGD_ENTRY;}extern inline int pgd_none(pgd_t pgd){	return pgd_val(pgd) & _PGD_ENTRY_INV;}extern inline int pgd_bad(pgd_t pgd){	return (pgd_val(pgd) & (~PAGE_MASK & ~_PGD_ENTRY_INV)) != _PGD_ENTRY;}extern inline int pmd_present(pmd_t pmd){	return (pmd_val(pmd) & ~PAGE_MASK) == _PMD_ENTRY;}extern inline int pmd_none(pmd_t pmd){

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -