⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 pci_iommu.c

📁 Linux内核源代码 为压缩文件 是<<Linux内核>>一书中的源代码
💻 C
📖 第 1 页 / 共 2 页
字号:
/* *	linux/arch/alpha/kernel/pci_iommu.c */#include <linux/kernel.h>#include <linux/mm.h>#include <linux/pci.h>#include <linux/slab.h>#include <linux/bootmem.h>#include <asm/io.h>#include <asm/hwrpb.h>#include "proto.h"#include "pci_impl.h"#define DEBUG_ALLOC 0#if DEBUG_ALLOC > 0# define DBGA(args...)		printk(KERN_DEBUG ##args)#else# define DBGA(args...)#endif#if DEBUG_ALLOC > 1# define DBGA2(args...)		printk(KERN_DEBUG ##args)#else# define DBGA2(args...)#endif#define DEBUG_NODIRECT 0static inline unsigned longmk_iommu_pte(unsigned long paddr){	return (paddr >> (PAGE_SHIFT-1)) | 1;}static inline longcalc_npages(long bytes){	return (bytes + PAGE_SIZE - 1) >> PAGE_SHIFT;}struct pci_iommu_arena *iommu_arena_new(struct pci_controler *hose, dma_addr_t base,		unsigned long window_size, unsigned long align){	unsigned long mem_size;	struct pci_iommu_arena *arena;	mem_size = window_size / (PAGE_SIZE / sizeof(unsigned long));	/* Note that the TLB lookup logic uses bitwise concatenation,	   not addition, so the required arena alignment is based on	   the size of the window.  Retain the align parameter so that	   particular systems can over-align the arena.  */	if (align < mem_size)		align = mem_size;	arena = alloc_bootmem(sizeof(*arena));	arena->ptes = __alloc_bootmem(mem_size, align, 0);	spin_lock_init(&arena->lock);	arena->hose = hose;	arena->dma_base = base;	arena->size = window_size;	arena->next_entry = 0;	/* Align allocations to a multiple of a page size.  Not needed	   unless there are chip bugs.  */	arena->align_entry = 1;	return arena;}longiommu_arena_alloc(struct pci_iommu_arena *arena, long n){	unsigned long flags;	unsigned long *ptes;	long i, p, nent, mask;	spin_lock_irqsave(&arena->lock, flags);	/* Search forward for the first sequence of N empty ptes.  */	ptes = arena->ptes;	nent = arena->size >> PAGE_SHIFT;	mask = arena->align_entry - 1;	p = (arena->next_entry + mask) & ~mask;	i = 0;	while (i < n && p+i < nent) {		if (ptes[p+i])			p = (p + i + 1 + mask) & ~mask, i = 0;		else			i = i + 1;	}	if (i < n) {                /* Reached the end.  Flush the TLB and restart the                   search from the beginning.  */		alpha_mv.mv_pci_tbi(arena->hose, 0, -1);		p = 0, i = 0;		while (i < n && p+i < nent) {			if (ptes[p+i])				p = (p + i + 1 + mask) & ~mask, i = 0;			else				i = i + 1;		}		if (i < n) {			spin_unlock_irqrestore(&arena->lock, flags);			return -1;		}	}	/* Success.  Mark them all in use, ie not zero.  Typically	   bit zero is the valid bit, so write ~1 into everything.	   The chip specific bits will fill this in with something	   kosher when we return.  */	for (i = 0; i < n; ++i)		ptes[p+i] = ~1UL;	arena->next_entry = p + n;	spin_unlock_irqrestore(&arena->lock, flags);	return p;}static voidiommu_arena_free(struct pci_iommu_arena *arena, long ofs, long n){	unsigned long *p;	long i;	p = arena->ptes + ofs;	for (i = 0; i < n; ++i)		p[i] = 0;}/* Map a single buffer of the indicated size for PCI DMA in streaming   mode.  The 32-bit PCI bus mastering address to use is returned.   Once the device is given the dma address, the device owns this memory   until either pci_unmap_single or pci_dma_sync_single is performed.  */dma_addr_tpci_map_single(struct pci_dev *pdev, void *cpu_addr, long size, int direction){	struct pci_controler *hose = pdev ? pdev->sysdata : pci_isa_hose;	dma_addr_t max_dma = pdev ? pdev->dma_mask : 0x00ffffff;	struct pci_iommu_arena *arena;	long npages, dma_ofs, i;	unsigned long paddr;	dma_addr_t ret;	if (direction == PCI_DMA_NONE)		BUG();	paddr = virt_to_phys(cpu_addr);#if !DEBUG_NODIRECT	/* First check to see if we can use the direct map window.  */	if (paddr + size + __direct_map_base - 1 <= max_dma	    && paddr + size <= __direct_map_size) {		ret = paddr + __direct_map_base;		DBGA2("pci_map_single: [%p,%lx] -> direct %x from %p\n",		      cpu_addr, size, ret, __builtin_return_address(0));		return ret;	}#endif	/* If the machine doesn't define a pci_tbi routine, we have to	   assume it doesn't support sg mapping.  */	if (! alpha_mv.mv_pci_tbi) {		printk(KERN_INFO "pci_map_single failed: no hw sg\n");		return 0;	}			arena = hose->sg_pci;	if (!arena || arena->dma_base + arena->size > max_dma)		arena = hose->sg_isa;	npages = calc_npages((paddr & ~PAGE_MASK) + size);	dma_ofs = iommu_arena_alloc(arena, npages);	if (dma_ofs < 0) {		printk(KERN_INFO "pci_map_single failed: "		       "could not allocate dma page tables\n");		return 0;	}	paddr &= PAGE_MASK;	for (i = 0; i < npages; ++i, paddr += PAGE_SIZE)		arena->ptes[i + dma_ofs] = mk_iommu_pte(paddr);	ret = arena->dma_base + dma_ofs * PAGE_SIZE;	ret += (unsigned long)cpu_addr & ~PAGE_MASK;	DBGA("pci_map_single: [%p,%lx] np %ld -> sg %x from %p\n",	     cpu_addr, size, npages, ret, __builtin_return_address(0));	return ret;}/* Unmap a single streaming mode DMA translation.  The DMA_ADDR and   SIZE must match what was provided for in a previous pci_map_single   call.  All other usages are undefined.  After this call, reads by   the cpu to the buffer are guarenteed to see whatever the device   wrote there.  */voidpci_unmap_single(struct pci_dev *pdev, dma_addr_t dma_addr, long size,		 int direction){	struct pci_controler *hose = pdev ? pdev->sysdata : pci_isa_hose;	struct pci_iommu_arena *arena;	long dma_ofs, npages;	if (direction == PCI_DMA_NONE)		BUG();#if !DEBUG_NODIRECT	if (dma_addr >= __direct_map_base	    && dma_addr < __direct_map_base + __direct_map_size) {		/* Nothing to do.  */		DBGA2("pci_unmap_single: direct [%x,%lx] from %p\n",		      dma_addr, size, __builtin_return_address(0));		return;	}#endif	arena = hose->sg_pci;	if (!arena || dma_addr < arena->dma_base)		arena = hose->sg_isa;	dma_ofs = (dma_addr - arena->dma_base) >> PAGE_SHIFT;	if (dma_ofs * PAGE_SIZE >= arena->size) {		printk(KERN_ERR "Bogus pci_unmap_single: dma_addr %x "		       " base %x size %x\n", dma_addr, arena->dma_base,		       arena->size);		return;		BUG();	}	npages = calc_npages((dma_addr & ~PAGE_MASK) + size);	iommu_arena_free(arena, dma_ofs, npages);        /*	   If we're freeing ptes above the `next_entry' pointer (they           may have snuck back into the TLB since the last wrap flush),           we need to flush the TLB before reallocating the latter.	*/	if (dma_ofs >= arena->next_entry)		alpha_mv.mv_pci_tbi(hose, dma_addr, dma_addr + size - 1);	DBGA("pci_unmap_single: sg [%x,%lx] np %ld from %p\n",	     dma_addr, size, npages, __builtin_return_address(0));}/* Allocate and map kernel buffer using consistent mode DMA for PCI   device.  Returns non-NULL cpu-view pointer to the buffer if   successful and sets *DMA_ADDRP to the pci side dma address as well,   else DMA_ADDRP is undefined.  */void *pci_alloc_consistent(struct pci_dev *pdev, long size, dma_addr_t *dma_addrp){	void *cpu_addr;	long order = get_order(size);	cpu_addr = (void *)__get_free_pages(GFP_ATOMIC, order);	if (! cpu_addr) {		printk(KERN_INFO "pci_alloc_consistent: "		       "get_free_pages failed from %p\n",			__builtin_return_address(0));		/* ??? Really atomic allocation?  Otherwise we could play		   with vmalloc and sg if we can't find contiguous memory.  */		return NULL;	}	memset(cpu_addr, 0, size);	*dma_addrp = pci_map_single(pdev, cpu_addr, size,				    PCI_DMA_BIDIRECTIONAL);	if (*dma_addrp == 0) {		free_pages((unsigned long)cpu_addr, order);		return NULL;	}			DBGA2("pci_alloc_consistent: %lx -> [%p,%x] from %p\n",	      size, cpu_addr, *dma_addrp, __builtin_return_address(0));	return cpu_addr;}/* Free and unmap a consistent DMA buffer.  CPU_ADDR and DMA_ADDR must   be values that were returned from pci_alloc_consistent.  SIZE must   be the same as what as passed into pci_alloc_consistent.   References to the memory and mappings assosciated with CPU_ADDR or   DMA_ADDR past this call are illegal.  */voidpci_free_consistent(struct pci_dev *pdev, long size, void *cpu_addr,		    dma_addr_t dma_addr){

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -