⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 mga_dma.c

📁 linux-2.6.15.6
💻 C
📖 第 1 页 / 共 2 页
字号:
 * decrease the request size by a single page each pass through the loop. * * \todo * Determine whether the maximum address passed to drm_pci_alloc is correct. * The same goes for drm_addbufs_pci. * * \sa mga_do_dma_bootstrap, mga_do_agp_dma_bootstrap */static int mga_do_pci_dma_bootstrap(drm_device_t * dev,				    drm_mga_dma_bootstrap_t * dma_bs){	drm_mga_private_t *const dev_priv =	    (drm_mga_private_t *) dev->dev_private;	unsigned int warp_size = mga_warp_microcode_size(dev_priv);	unsigned int primary_size;	unsigned int bin_count;	int err;	drm_buf_desc_t req;	if (dev->dma == NULL) {		DRM_ERROR("dev->dma is NULL\n");		return DRM_ERR(EFAULT);	}	/* Make drm_addbufs happy by not trying to create a mapping for less	 * than a page.	 */	if (warp_size < PAGE_SIZE)		warp_size = PAGE_SIZE;	/* The proper alignment is 0x100 for this mapping */	err = drm_addmap(dev, 0, warp_size, _DRM_CONSISTENT,			 _DRM_READ_ONLY, &dev_priv->warp);	if (err != 0) {		DRM_ERROR("Unable to create mapping for WARP microcode\n");		return err;	}	/* Other than the bottom two bits being used to encode other	 * information, there don't appear to be any restrictions on the	 * alignment of the primary or secondary DMA buffers.	 */	for (primary_size = dma_bs->primary_size; primary_size != 0;	     primary_size >>= 1) {		/* The proper alignment for this mapping is 0x04 */		err = drm_addmap(dev, 0, primary_size, _DRM_CONSISTENT,				 _DRM_READ_ONLY, &dev_priv->primary);		if (!err)			break;	}	if (err != 0) {		DRM_ERROR("Unable to allocate primary DMA region\n");		return DRM_ERR(ENOMEM);	}	if (dev_priv->primary->size != dma_bs->primary_size) {		DRM_INFO("Primary DMA buffer size reduced from %u to %u.\n",			 dma_bs->primary_size,			 (unsigned)dev_priv->primary->size);		dma_bs->primary_size = dev_priv->primary->size;	}	for (bin_count = dma_bs->secondary_bin_count; bin_count > 0;	     bin_count--) {		(void)memset(&req, 0, sizeof(req));		req.count = bin_count;		req.size = dma_bs->secondary_bin_size;		err = drm_addbufs_pci(dev, &req);		if (!err) {			break;		}	}	if (bin_count == 0) {		DRM_ERROR("Unable to add secondary DMA buffers\n");		return err;	}	if (bin_count != dma_bs->secondary_bin_count) {		DRM_INFO("Secondary PCI DMA buffer bin count reduced from %u "			 "to %u.\n", dma_bs->secondary_bin_count, bin_count);		dma_bs->secondary_bin_count = bin_count;	}	dev_priv->dma_access = 0;	dev_priv->wagp_enable = 0;	dma_bs->agp_mode = 0;	DRM_INFO("Initialized card for PCI DMA.\n");	return 0;}static int mga_do_dma_bootstrap(drm_device_t * dev,				drm_mga_dma_bootstrap_t * dma_bs){	const int is_agp = (dma_bs->agp_mode != 0) && drm_device_is_agp(dev);	int err;	drm_mga_private_t *const dev_priv =	    (drm_mga_private_t *) dev->dev_private;	dev_priv->used_new_dma_init = 1;	/* The first steps are the same for both PCI and AGP based DMA.  Map	 * the cards MMIO registers and map a status page.	 */	err = drm_addmap(dev, dev_priv->mmio_base, dev_priv->mmio_size,			 _DRM_REGISTERS, _DRM_READ_ONLY, &dev_priv->mmio);	if (err) {		DRM_ERROR("Unable to map MMIO region\n");		return err;	}	err = drm_addmap(dev, 0, SAREA_MAX, _DRM_SHM,			 _DRM_READ_ONLY | _DRM_LOCKED | _DRM_KERNEL,			 &dev_priv->status);	if (err) {		DRM_ERROR("Unable to map status region\n");		return err;	}	/* The DMA initialization procedure is slightly different for PCI and	 * AGP cards.  AGP cards just allocate a large block of AGP memory and	 * carve off portions of it for internal uses.  The remaining memory	 * is returned to user-mode to be used for AGP textures.	 */	if (is_agp) {		err = mga_do_agp_dma_bootstrap(dev, dma_bs);	}	/* If we attempted to initialize the card for AGP DMA but failed,	 * clean-up any mess that may have been created.	 */	if (err) {		mga_do_cleanup_dma(dev);	}	/* Not only do we want to try and initialized PCI cards for PCI DMA,	 * but we also try to initialized AGP cards that could not be	 * initialized for AGP DMA.  This covers the case where we have an AGP	 * card in a system with an unsupported AGP chipset.  In that case the	 * card will be detected as AGP, but we won't be able to allocate any	 * AGP memory, etc.	 */	if (!is_agp || err) {		err = mga_do_pci_dma_bootstrap(dev, dma_bs);	}	return err;}int mga_dma_bootstrap(DRM_IOCTL_ARGS){	DRM_DEVICE;	drm_mga_dma_bootstrap_t bootstrap;	int err;	DRM_COPY_FROM_USER_IOCTL(bootstrap,				 (drm_mga_dma_bootstrap_t __user *) data,				 sizeof(bootstrap));	err = mga_do_dma_bootstrap(dev, &bootstrap);	if (!err) {		static const int modes[] = { 0, 1, 2, 2, 4, 4, 4, 4 };		const drm_mga_private_t *const dev_priv =		    (drm_mga_private_t *) dev->dev_private;		if (dev_priv->agp_textures != NULL) {			bootstrap.texture_handle =			    dev_priv->agp_textures->offset;			bootstrap.texture_size = dev_priv->agp_textures->size;		} else {			bootstrap.texture_handle = 0;			bootstrap.texture_size = 0;		}		bootstrap.agp_mode = modes[bootstrap.agp_mode & 0x07];		if (DRM_COPY_TO_USER((void __user *)data, &bootstrap,				     sizeof(bootstrap))) {			err = DRM_ERR(EFAULT);		}	} else {		mga_do_cleanup_dma(dev);	}	return err;}static int mga_do_init_dma(drm_device_t * dev, drm_mga_init_t * init){	drm_mga_private_t *dev_priv;	int ret;	DRM_DEBUG("\n");	dev_priv = dev->dev_private;	if (init->sgram) {		dev_priv->clear_cmd = MGA_DWGCTL_CLEAR | MGA_ATYPE_BLK;	} else {		dev_priv->clear_cmd = MGA_DWGCTL_CLEAR | MGA_ATYPE_RSTR;	}	dev_priv->maccess = init->maccess;	dev_priv->fb_cpp = init->fb_cpp;	dev_priv->front_offset = init->front_offset;	dev_priv->front_pitch = init->front_pitch;	dev_priv->back_offset = init->back_offset;	dev_priv->back_pitch = init->back_pitch;	dev_priv->depth_cpp = init->depth_cpp;	dev_priv->depth_offset = init->depth_offset;	dev_priv->depth_pitch = init->depth_pitch;	/* FIXME: Need to support AGP textures...	 */	dev_priv->texture_offset = init->texture_offset[0];	dev_priv->texture_size = init->texture_size[0];	DRM_GETSAREA();	if (!dev_priv->sarea) {		DRM_ERROR("failed to find sarea!\n");		return DRM_ERR(EINVAL);	}	if (!dev_priv->used_new_dma_init) {		dev_priv->dma_access = MGA_PAGPXFER;		dev_priv->wagp_enable = MGA_WAGP_ENABLE;		dev_priv->status = drm_core_findmap(dev, init->status_offset);		if (!dev_priv->status) {			DRM_ERROR("failed to find status page!\n");			return DRM_ERR(EINVAL);		}		dev_priv->mmio = drm_core_findmap(dev, init->mmio_offset);		if (!dev_priv->mmio) {			DRM_ERROR("failed to find mmio region!\n");			return DRM_ERR(EINVAL);		}		dev_priv->warp = drm_core_findmap(dev, init->warp_offset);		if (!dev_priv->warp) {			DRM_ERROR("failed to find warp microcode region!\n");			return DRM_ERR(EINVAL);		}		dev_priv->primary = drm_core_findmap(dev, init->primary_offset);		if (!dev_priv->primary) {			DRM_ERROR("failed to find primary dma region!\n");			return DRM_ERR(EINVAL);		}		dev->agp_buffer_token = init->buffers_offset;		dev->agp_buffer_map =		    drm_core_findmap(dev, init->buffers_offset);		if (!dev->agp_buffer_map) {			DRM_ERROR("failed to find dma buffer region!\n");			return DRM_ERR(EINVAL);		}		drm_core_ioremap(dev_priv->warp, dev);		drm_core_ioremap(dev_priv->primary, dev);		drm_core_ioremap(dev->agp_buffer_map, dev);	}	dev_priv->sarea_priv =	    (drm_mga_sarea_t *) ((u8 *) dev_priv->sarea->handle +				 init->sarea_priv_offset);	if (!dev_priv->warp->handle ||	    !dev_priv->primary->handle ||	    ((dev_priv->dma_access != 0) &&	     ((dev->agp_buffer_map == NULL) ||	      (dev->agp_buffer_map->handle == NULL)))) {		DRM_ERROR("failed to ioremap agp regions!\n");		return DRM_ERR(ENOMEM);	}	ret = mga_warp_install_microcode(dev_priv);	if (ret < 0) {		DRM_ERROR("failed to install WARP ucode!\n");		return ret;	}	ret = mga_warp_init(dev_priv);	if (ret < 0) {		DRM_ERROR("failed to init WARP engine!\n");		return ret;	}	dev_priv->prim.status = (u32 *) dev_priv->status->handle;	mga_do_wait_for_idle(dev_priv);	/* Init the primary DMA registers.	 */	MGA_WRITE(MGA_PRIMADDRESS, dev_priv->primary->offset | MGA_DMA_GENERAL);#if 0	MGA_WRITE(MGA_PRIMPTR, virt_to_bus((void *)dev_priv->prim.status) | MGA_PRIMPTREN0 |	/* Soft trap, SECEND, SETUPEND */		  MGA_PRIMPTREN1);	/* DWGSYNC */#endif	dev_priv->prim.start = (u8 *) dev_priv->primary->handle;	dev_priv->prim.end = ((u8 *) dev_priv->primary->handle			      + dev_priv->primary->size);	dev_priv->prim.size = dev_priv->primary->size;	dev_priv->prim.tail = 0;	dev_priv->prim.space = dev_priv->prim.size;	dev_priv->prim.wrapped = 0;	dev_priv->prim.last_flush = 0;	dev_priv->prim.last_wrap = 0;	dev_priv->prim.high_mark = 256 * DMA_BLOCK_SIZE;	dev_priv->prim.status[0] = dev_priv->primary->offset;	dev_priv->prim.status[1] = 0;	dev_priv->sarea_priv->last_wrap = 0;	dev_priv->sarea_priv->last_frame.head = 0;	dev_priv->sarea_priv->last_frame.wrap = 0;	if (mga_freelist_init(dev, dev_priv) < 0) {		DRM_ERROR("could not initialize freelist\n");		return DRM_ERR(ENOMEM);	}	return 0;}static int mga_do_cleanup_dma(drm_device_t * dev){	int err = 0;	DRM_DEBUG("\n");	/* Make sure interrupts are disabled here because the uninstall ioctl	 * may not have been called from userspace and after dev_private	 * is freed, it's too late.	 */	if (dev->irq_enabled)		drm_irq_uninstall(dev);	if (dev->dev_private) {		drm_mga_private_t *dev_priv = dev->dev_private;		if ((dev_priv->warp != NULL)		    && (dev_priv->warp->type != _DRM_CONSISTENT))			drm_core_ioremapfree(dev_priv->warp, dev);		if ((dev_priv->primary != NULL)		    && (dev_priv->primary->type != _DRM_CONSISTENT))			drm_core_ioremapfree(dev_priv->primary, dev);		if (dev->agp_buffer_map != NULL)			drm_core_ioremapfree(dev->agp_buffer_map, dev);		if (dev_priv->used_new_dma_init) {#if __OS_HAS_AGP			if (dev_priv->agp_mem != NULL) {				dev_priv->agp_textures = NULL;				drm_unbind_agp(dev_priv->agp_mem);				drm_free_agp(dev_priv->agp_mem,					     dev_priv->agp_pages);				dev_priv->agp_pages = 0;				dev_priv->agp_mem = NULL;			}			if ((dev->agp != NULL) && dev->agp->acquired) {				err = drm_agp_release(dev);			}#endif			dev_priv->used_new_dma_init = 0;		}		dev_priv->warp = NULL;		dev_priv->primary = NULL;		dev_priv->mmio = NULL;		dev_priv->status = NULL;		dev_priv->sarea = NULL;		dev_priv->sarea_priv = NULL;		dev->agp_buffer_map = NULL;		memset(&dev_priv->prim, 0, sizeof(dev_priv->prim));		dev_priv->warp_pipe = 0;		memset(dev_priv->warp_pipe_phys, 0,		       sizeof(dev_priv->warp_pipe_phys));		if (dev_priv->head != NULL) {			mga_freelist_cleanup(dev);		}	}	return err;}int mga_dma_init(DRM_IOCTL_ARGS){	DRM_DEVICE;	drm_mga_init_t init;	int err;	LOCK_TEST_WITH_RETURN(dev, filp);	DRM_COPY_FROM_USER_IOCTL(init, (drm_mga_init_t __user *) data,				 sizeof(init));	switch (init.func) {	case MGA_INIT_DMA:		err = mga_do_init_dma(dev, &init);		if (err) {			(void)mga_do_cleanup_dma(dev);		}		return err;	case MGA_CLEANUP_DMA:		return mga_do_cleanup_dma(dev);	}	return DRM_ERR(EINVAL);}/* ================================================================ * Primary DMA stream management */int mga_dma_flush(DRM_IOCTL_ARGS){	DRM_DEVICE;	drm_mga_private_t *dev_priv = (drm_mga_private_t *) dev->dev_private;	drm_lock_t lock;	LOCK_TEST_WITH_RETURN(dev, filp);	DRM_COPY_FROM_USER_IOCTL(lock, (drm_lock_t __user *) data,				 sizeof(lock));	DRM_DEBUG("%s%s%s\n",		  (lock.flags & _DRM_LOCK_FLUSH) ? "flush, " : "",		  (lock.flags & _DRM_LOCK_FLUSH_ALL) ? "flush all, " : "",		  (lock.flags & _DRM_LOCK_QUIESCENT) ? "idle, " : "");	WRAP_WAIT_WITH_RETURN(dev_priv);	if (lock.flags & (_DRM_LOCK_FLUSH | _DRM_LOCK_FLUSH_ALL)) {		mga_do_dma_flush(dev_priv);	}	if (lock.flags & _DRM_LOCK_QUIESCENT) {#if MGA_DMA_DEBUG		int ret = mga_do_wait_for_idle(dev_priv);		if (ret < 0)			DRM_INFO("%s: -EBUSY\n", __FUNCTION__);		return ret;#else		return mga_do_wait_for_idle(dev_priv);#endif	} else {		return 0;	}}int mga_dma_reset(DRM_IOCTL_ARGS){	DRM_DEVICE;	drm_mga_private_t *dev_priv = (drm_mga_private_t *) dev->dev_private;	LOCK_TEST_WITH_RETURN(dev, filp);	return mga_do_dma_reset(dev_priv);}/* ================================================================ * DMA buffer management */static int mga_dma_get_buffers(DRMFILE filp, drm_device_t * dev, drm_dma_t * d){	drm_buf_t *buf;	int i;	for (i = d->granted_count; i < d->request_count; i++) {		buf = mga_freelist_get(dev);		if (!buf)			return DRM_ERR(EAGAIN);		buf->filp = filp;		if (DRM_COPY_TO_USER(&d->request_indices[i],				     &buf->idx, sizeof(buf->idx)))			return DRM_ERR(EFAULT);		if (DRM_COPY_TO_USER(&d->request_sizes[i],				     &buf->total, sizeof(buf->total)))			return DRM_ERR(EFAULT);		d->granted_count++;	}	return 0;}int mga_dma_buffers(DRM_IOCTL_ARGS){	DRM_DEVICE;	drm_device_dma_t *dma = dev->dma;	drm_mga_private_t *dev_priv = (drm_mga_private_t *) dev->dev_private;	drm_dma_t __user *argp = (void __user *)data;	drm_dma_t d;	int ret = 0;	LOCK_TEST_WITH_RETURN(dev, filp);	DRM_COPY_FROM_USER_IOCTL(d, argp, sizeof(d));	/* Please don't send us buffers.	 */	if (d.send_count != 0) {		DRM_ERROR("Process %d trying to send %d buffers via drmDMA\n",			  DRM_CURRENTPID, d.send_count);		return DRM_ERR(EINVAL);	}	/* We'll send you buffers.	 */	if (d.request_count < 0 || d.request_count > dma->buf_count) {		DRM_ERROR("Process %d trying to get %d buffers (of %d max)\n",			  DRM_CURRENTPID, d.request_count, dma->buf_count);		return DRM_ERR(EINVAL);	}	WRAP_TEST_WITH_RETURN(dev_priv);	d.granted_count = 0;	if (d.request_count) {		ret = mga_dma_get_buffers(filp, dev, &d);	}	DRM_COPY_TO_USER_IOCTL(argp, d, sizeof(d));	return ret;}/** * Called just before the module is unloaded. */int mga_driver_postcleanup(drm_device_t * dev){	drm_free(dev->dev_private, sizeof(drm_mga_private_t), DRM_MEM_DRIVER);	dev->dev_private = NULL;	return 0;}/** * Called when the last opener of the device is closed. */void mga_driver_pretakedown(drm_device_t * dev){	mga_do_cleanup_dma(dev);}int mga_driver_dma_quiescent(drm_device_t * dev){	drm_mga_private_t *dev_priv = dev->dev_private;	return mga_do_wait_for_idle(dev_priv);}

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -