| /* drm_drv.h -- Generic driver template -*- linux-c -*- |
| * Created: Thu Nov 23 03:10:50 2000 by gareth@valinux.com |
| * |
| * Copyright 1999, 2000 Precision Insight, Inc., Cedar Park, Texas. |
| * Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California. |
| * All Rights Reserved. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice (including the next |
| * paragraph) shall be included in all copies or substantial portions of the |
| * Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR |
| * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
| * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
| * OTHER DEALINGS IN THE SOFTWARE. |
| * |
| * Authors: |
| * Rickard E. (Rik) Faith <faith@valinux.com> |
| * Gareth Hughes <gareth@valinux.com> |
| */ |
| |
| /* |
| * To use this template, you must at least define the following (samples |
| * given for the MGA driver): |
| * |
| * #define DRIVER_AUTHOR "VA Linux Systems, Inc." |
| * |
| * #define DRIVER_NAME "mga" |
| * #define DRIVER_DESC "Matrox G200/G400" |
| * #define DRIVER_DATE "20001127" |
| * |
| * #define DRIVER_MAJOR 2 |
| * #define DRIVER_MINOR 0 |
| * #define DRIVER_PATCHLEVEL 2 |
| * |
| * #define DRIVER_IOCTL_COUNT DRM_ARRAY_SIZE( mga_ioctls ) |
| * |
| * #define DRM(x) mga_##x |
| */ |
| |
| #ifndef __MUST_HAVE_AGP |
| #define __MUST_HAVE_AGP 0 |
| #endif |
| #ifndef __HAVE_CTX_BITMAP |
| #define __HAVE_CTX_BITMAP 0 |
| #endif |
| #ifndef __HAVE_DMA_IRQ |
| #define __HAVE_DMA_IRQ 0 |
| #endif |
| #ifndef __HAVE_DMA_QUEUE |
| #define __HAVE_DMA_QUEUE 0 |
| #endif |
| #ifndef __HAVE_MULTIPLE_DMA_QUEUES |
| #define __HAVE_MULTIPLE_DMA_QUEUES 0 |
| #endif |
| #ifndef __HAVE_DMA_SCHEDULE |
| #define __HAVE_DMA_SCHEDULE 0 |
| #endif |
| #ifndef __HAVE_DMA_FLUSH |
| #define __HAVE_DMA_FLUSH 0 |
| #endif |
| #ifndef __HAVE_DMA_READY |
| #define __HAVE_DMA_READY 0 |
| #endif |
| #ifndef __HAVE_DMA_QUIESCENT |
| #define __HAVE_DMA_QUIESCENT 0 |
| #endif |
| #ifndef __HAVE_RELEASE |
| #define __HAVE_RELEASE 0 |
| #endif |
| #ifndef __HAVE_COUNTERS |
| #define __HAVE_COUNTERS 0 |
| #endif |
| #ifndef __HAVE_SG |
| #define __HAVE_SG 0 |
| #endif |
| #ifndef __HAVE_KERNEL_CTX_SWITCH |
| #define __HAVE_KERNEL_CTX_SWITCH 0 |
| #endif |
| |
| #ifndef DRIVER_PREINIT |
| #define DRIVER_PREINIT() |
| #endif |
| #ifndef DRIVER_POSTINIT |
| #define DRIVER_POSTINIT() |
| #endif |
| #ifndef DRIVER_PRERELEASE |
| #define DRIVER_PRERELEASE() |
| #endif |
| #ifndef DRIVER_PRETAKEDOWN |
| #define DRIVER_PRETAKEDOWN() |
| #endif |
| #ifndef DRIVER_POSTCLEANUP |
| #define DRIVER_POSTCLEANUP() |
| #endif |
| #ifndef DRIVER_PRESETUP |
| #define DRIVER_PRESETUP() |
| #endif |
| #ifndef DRIVER_POSTSETUP |
| #define DRIVER_POSTSETUP() |
| #endif |
| #ifndef DRIVER_IOCTLS |
| #define DRIVER_IOCTLS |
| #endif |
| #ifndef DRIVER_FOPS |
| #define DRIVER_FOPS \ |
| static struct file_operations DRM(fops) = { \ |
| .owner = THIS_MODULE, \ |
| .open = DRM(open), \ |
| .flush = DRM(flush), \ |
| .release = DRM(release), \ |
| .ioctl = DRM(ioctl), \ |
| .mmap = DRM(mmap), \ |
| .read = DRM(read), \ |
| .fasync = DRM(fasync), \ |
| .poll = DRM(poll), \ |
| } |
| #endif |
| |
| #ifndef MODULE |
| /* DRM(options) is called by the kernel to parse command-line options |
| * passed via the boot-loader (e.g., LILO). It calls the insmod option |
| * routine, drm_parse_drm. |
| */ |
| /* Use an additional macro to avoid preprocessor troubles */ |
| #define DRM_OPTIONS_FUNC DRM(options) |
| static int __init DRM(options)( char *str ) |
| { |
| DRM(parse_options)( str ); |
| return 1; |
| } |
| |
| __setup( DRIVER_NAME "=", DRM_OPTIONS_FUNC ); |
| #undef DRM_OPTIONS_FUNC |
| #endif |
| |
| /* |
| * The default number of instances (minor numbers) to initialize. |
| */ |
| #ifndef DRIVER_NUM_CARDS |
| #define DRIVER_NUM_CARDS 1 |
| #endif |
| |
| static drm_device_t *DRM(device); |
| static int *DRM(minor); |
| static int DRM(numdevs) = 0; |
| |
| DRIVER_FOPS; |
| |
| static drm_ioctl_desc_t DRM(ioctls)[] = { |
| [DRM_IOCTL_NR(DRM_IOCTL_VERSION)] = { DRM(version), 0, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_UNIQUE)] = { DRM(getunique), 0, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_MAGIC)] = { DRM(getmagic), 0, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_IRQ_BUSID)] = { DRM(irq_busid), 0, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_MAP)] = { DRM(getmap), 0, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_CLIENT)] = { DRM(getclient), 0, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_STATS)] = { DRM(getstats), 0, 0 }, |
| |
| [DRM_IOCTL_NR(DRM_IOCTL_SET_UNIQUE)] = { DRM(setunique), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_BLOCK)] = { DRM(block), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_UNBLOCK)] = { DRM(unblock), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AUTH_MAGIC)] = { DRM(authmagic), 1, 1 }, |
| |
| [DRM_IOCTL_NR(DRM_IOCTL_ADD_MAP)] = { DRM(addmap), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_RM_MAP)] = { DRM(rmmap), 1, 0 }, |
| |
| #if __HAVE_CTX_BITMAP |
| [DRM_IOCTL_NR(DRM_IOCTL_SET_SAREA_CTX)] = { DRM(setsareactx), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_SAREA_CTX)] = { DRM(getsareactx), 1, 0 }, |
| #endif |
| |
| [DRM_IOCTL_NR(DRM_IOCTL_ADD_CTX)] = { DRM(addctx), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_RM_CTX)] = { DRM(rmctx), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_MOD_CTX)] = { DRM(modctx), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_GET_CTX)] = { DRM(getctx), 1, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_SWITCH_CTX)] = { DRM(switchctx), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_NEW_CTX)] = { DRM(newctx), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_RES_CTX)] = { DRM(resctx), 1, 0 }, |
| |
| [DRM_IOCTL_NR(DRM_IOCTL_ADD_DRAW)] = { DRM(adddraw), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_RM_DRAW)] = { DRM(rmdraw), 1, 1 }, |
| |
| [DRM_IOCTL_NR(DRM_IOCTL_LOCK)] = { DRM(lock), 1, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_UNLOCK)] = { DRM(unlock), 1, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_FINISH)] = { DRM(finish), 1, 0 }, |
| |
| #if __HAVE_DMA |
| [DRM_IOCTL_NR(DRM_IOCTL_ADD_BUFS)] = { DRM(addbufs), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_MARK_BUFS)] = { DRM(markbufs), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_INFO_BUFS)] = { DRM(infobufs), 1, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_MAP_BUFS)] = { DRM(mapbufs), 1, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_FREE_BUFS)] = { DRM(freebufs), 1, 0 }, |
| |
| /* The DRM_IOCTL_DMA ioctl should be defined by the driver. |
| */ |
| [DRM_IOCTL_NR(DRM_IOCTL_CONTROL)] = { DRM(control), 1, 1 }, |
| #endif |
| |
| #if __REALLY_HAVE_AGP |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_ACQUIRE)] = { DRM(agp_acquire), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_RELEASE)] = { DRM(agp_release), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_ENABLE)] = { DRM(agp_enable), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_INFO)] = { DRM(agp_info), 1, 0 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_ALLOC)] = { DRM(agp_alloc), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_FREE)] = { DRM(agp_free), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_BIND)] = { DRM(agp_bind), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_AGP_UNBIND)] = { DRM(agp_unbind), 1, 1 }, |
| #endif |
| |
| #if __HAVE_SG |
| [DRM_IOCTL_NR(DRM_IOCTL_SG_ALLOC)] = { DRM(sg_alloc), 1, 1 }, |
| [DRM_IOCTL_NR(DRM_IOCTL_SG_FREE)] = { DRM(sg_free), 1, 1 }, |
| #endif |
| |
| DRIVER_IOCTLS |
| }; |
| |
| #define DRIVER_IOCTL_COUNT DRM_ARRAY_SIZE( DRM(ioctls) ) |
| |
| #ifdef MODULE |
| static char *drm_opts = NULL; |
| #endif |
| |
| MODULE_AUTHOR( DRIVER_AUTHOR ); |
| MODULE_DESCRIPTION( DRIVER_DESC ); |
| MODULE_PARM( drm_opts, "s" ); |
| MODULE_LICENSE("GPL and additional rights"); |
| |
| static int DRM(setup)( drm_device_t *dev ) |
| { |
| int i; |
| |
| DRIVER_PRESETUP(); |
| atomic_set( &dev->ioctl_count, 0 ); |
| atomic_set( &dev->vma_count, 0 ); |
| dev->buf_use = 0; |
| atomic_set( &dev->buf_alloc, 0 ); |
| |
| #if __HAVE_DMA |
| i = DRM(dma_setup)( dev ); |
| if ( i < 0 ) |
| return i; |
| #endif |
| |
| dev->counters = 6 + __HAVE_COUNTERS; |
| dev->types[0] = _DRM_STAT_LOCK; |
| dev->types[1] = _DRM_STAT_OPENS; |
| dev->types[2] = _DRM_STAT_CLOSES; |
| dev->types[3] = _DRM_STAT_IOCTLS; |
| dev->types[4] = _DRM_STAT_LOCKS; |
| dev->types[5] = _DRM_STAT_UNLOCKS; |
| #ifdef __HAVE_COUNTER6 |
| dev->types[6] = __HAVE_COUNTER6; |
| #endif |
| #ifdef __HAVE_COUNTER7 |
| dev->types[7] = __HAVE_COUNTER7; |
| #endif |
| #ifdef __HAVE_COUNTER8 |
| dev->types[8] = __HAVE_COUNTER8; |
| #endif |
| #ifdef __HAVE_COUNTER9 |
| dev->types[9] = __HAVE_COUNTER9; |
| #endif |
| #ifdef __HAVE_COUNTER10 |
| dev->types[10] = __HAVE_COUNTER10; |
| #endif |
| #ifdef __HAVE_COUNTER11 |
| dev->types[11] = __HAVE_COUNTER11; |
| #endif |
| #ifdef __HAVE_COUNTER12 |
| dev->types[12] = __HAVE_COUNTER12; |
| #endif |
| #ifdef __HAVE_COUNTER13 |
| dev->types[13] = __HAVE_COUNTER13; |
| #endif |
| #ifdef __HAVE_COUNTER14 |
| dev->types[14] = __HAVE_COUNTER14; |
| #endif |
| #ifdef __HAVE_COUNTER15 |
| dev->types[14] = __HAVE_COUNTER14; |
| #endif |
| |
| for ( i = 0 ; i < DRM_ARRAY_SIZE(dev->counts) ; i++ ) |
| atomic_set( &dev->counts[i], 0 ); |
| |
| for ( i = 0 ; i < DRM_HASH_SIZE ; i++ ) { |
| dev->magiclist[i].head = NULL; |
| dev->magiclist[i].tail = NULL; |
| } |
| |
| dev->maplist = DRM(alloc)(sizeof(*dev->maplist), |
| DRM_MEM_MAPS); |
| if(dev->maplist == NULL) return -ENOMEM; |
| memset(dev->maplist, 0, sizeof(*dev->maplist)); |
| INIT_LIST_HEAD(&dev->maplist->head); |
| dev->map_count = 0; |
| |
| dev->vmalist = NULL; |
| dev->sigdata.lock = dev->lock.hw_lock = NULL; |
| init_waitqueue_head( &dev->lock.lock_queue ); |
| dev->queue_count = 0; |
| dev->queue_reserved = 0; |
| dev->queue_slots = 0; |
| dev->queuelist = NULL; |
| dev->irq = 0; |
| dev->context_flag = 0; |
| dev->interrupt_flag = 0; |
| dev->dma_flag = 0; |
| dev->last_context = 0; |
| dev->last_switch = 0; |
| dev->last_checked = 0; |
| init_timer( &dev->timer ); |
| init_waitqueue_head( &dev->context_wait ); |
| |
| dev->ctx_start = 0; |
| dev->lck_start = 0; |
| |
| dev->buf_rp = dev->buf; |
| dev->buf_wp = dev->buf; |
| dev->buf_end = dev->buf + DRM_BSZ; |
| dev->buf_async = NULL; |
| init_waitqueue_head( &dev->buf_readers ); |
| init_waitqueue_head( &dev->buf_writers ); |
| |
| DRM_DEBUG( "\n" ); |
| |
| /* The kernel's context could be created here, but is now created |
| * in drm_dma_enqueue. This is more resource-efficient for |
| * hardware that does not do DMA, but may mean that |
| * drm_select_queue fails between the time the interrupt is |
| * initialized and the time the queues are initialized. |
| */ |
| DRIVER_POSTSETUP(); |
| return 0; |
| } |
| |
| |
| static int DRM(takedown)( drm_device_t *dev ) |
| { |
| drm_magic_entry_t *pt, *next; |
| drm_map_t *map; |
| drm_map_list_t *r_list; |
| struct list_head *list, *list_next; |
| drm_vma_entry_t *vma, *vma_next; |
| int i; |
| |
| DRM_DEBUG( "\n" ); |
| |
| DRIVER_PRETAKEDOWN(); |
| #if __HAVE_DMA_IRQ |
| if ( dev->irq ) DRM(irq_uninstall)( dev ); |
| #endif |
| |
| down( &dev->struct_sem ); |
| del_timer( &dev->timer ); |
| |
| if ( dev->devname ) { |
| DRM(free)( dev->devname, strlen( dev->devname ) + 1, |
| DRM_MEM_DRIVER ); |
| dev->devname = NULL; |
| } |
| |
| if ( dev->unique ) { |
| DRM(free)( dev->unique, strlen( dev->unique ) + 1, |
| DRM_MEM_DRIVER ); |
| dev->unique = NULL; |
| dev->unique_len = 0; |
| } |
| /* Clear pid list */ |
| for ( i = 0 ; i < DRM_HASH_SIZE ; i++ ) { |
| for ( pt = dev->magiclist[i].head ; pt ; pt = next ) { |
| next = pt->next; |
| DRM(free)( pt, sizeof(*pt), DRM_MEM_MAGIC ); |
| } |
| dev->magiclist[i].head = dev->magiclist[i].tail = NULL; |
| } |
| |
| #if __REALLY_HAVE_AGP |
| /* Clear AGP information */ |
| if ( dev->agp ) { |
| drm_agp_mem_t *entry; |
| drm_agp_mem_t *nexte; |
| |
| /* Remove AGP resources, but leave dev->agp |
| intact until drv_cleanup is called. */ |
| for ( entry = dev->agp->memory ; entry ; entry = nexte ) { |
| nexte = entry->next; |
| if ( entry->bound ) DRM(unbind_agp)( entry->memory ); |
| DRM(free_agp)( entry->memory, entry->pages ); |
| DRM(free)( entry, sizeof(*entry), DRM_MEM_AGPLISTS ); |
| } |
| dev->agp->memory = NULL; |
| |
| if ( dev->agp->acquired ) DRM(agp_do_release)(); |
| |
| dev->agp->acquired = 0; |
| dev->agp->enabled = 0; |
| } |
| #endif |
| |
| /* Clear vma list (only built for debugging) */ |
| if ( dev->vmalist ) { |
| for ( vma = dev->vmalist ; vma ; vma = vma_next ) { |
| vma_next = vma->next; |
| DRM(free)( vma, sizeof(*vma), DRM_MEM_VMAS ); |
| } |
| dev->vmalist = NULL; |
| } |
| |
| if( dev->maplist ) { |
| for(list = dev->maplist->head.next; |
| list != &dev->maplist->head; |
| list = list_next) { |
| list_next = list->next; |
| r_list = (drm_map_list_t *)list; |
| map = r_list->map; |
| DRM(free)(r_list, sizeof(*r_list), DRM_MEM_MAPS); |
| if(!map) continue; |
| |
| switch ( map->type ) { |
| case _DRM_REGISTERS: |
| case _DRM_FRAME_BUFFER: |
| #if __REALLY_HAVE_MTRR |
| if ( map->mtrr >= 0 ) { |
| int retcode; |
| retcode = mtrr_del( map->mtrr, |
| map->offset, |
| map->size ); |
| DRM_DEBUG( "mtrr_del=%d\n", retcode ); |
| } |
| #endif |
| DRM(ioremapfree)( map->handle, map->size ); |
| break; |
| case _DRM_SHM: |
| vfree(map->handle); |
| break; |
| |
| case _DRM_AGP: |
| /* Do nothing here, because this is all |
| * handled in the AGP/GART driver. |
| */ |
| break; |
| case _DRM_SCATTER_GATHER: |
| /* Handle it, but do nothing, if HAVE_SG |
| * isn't defined. |
| */ |
| #if __HAVE_SG |
| if(dev->sg) { |
| DRM(sg_cleanup)(dev->sg); |
| dev->sg = NULL; |
| } |
| #endif |
| break; |
| } |
| DRM(free)(map, sizeof(*map), DRM_MEM_MAPS); |
| } |
| DRM(free)(dev->maplist, sizeof(*dev->maplist), DRM_MEM_MAPS); |
| dev->maplist = NULL; |
| } |
| |
| #if __HAVE_DMA_QUEUE || __HAVE_MULTIPLE_DMA_QUEUES |
| if ( dev->queuelist ) { |
| for ( i = 0 ; i < dev->queue_count ; i++ ) { |
| DRM(waitlist_destroy)( &dev->queuelist[i]->waitlist ); |
| if ( dev->queuelist[i] ) { |
| DRM(free)( dev->queuelist[i], |
| sizeof(*dev->queuelist[0]), |
| DRM_MEM_QUEUES ); |
| dev->queuelist[i] = NULL; |
| } |
| } |
| DRM(free)( dev->queuelist, |
| dev->queue_slots * sizeof(*dev->queuelist), |
| DRM_MEM_QUEUES ); |
| dev->queuelist = NULL; |
| } |
| dev->queue_count = 0; |
| #endif |
| |
| #if __HAVE_DMA |
| DRM(dma_takedown)( dev ); |
| #endif |
| if ( dev->lock.hw_lock ) { |
| dev->sigdata.lock = dev->lock.hw_lock = NULL; /* SHM removed */ |
| dev->lock.pid = 0; |
| wake_up_interruptible( &dev->lock.lock_queue ); |
| } |
| up( &dev->struct_sem ); |
| |
| return 0; |
| } |
| |
| /* |
| * Figure out how many instances to initialize. |
| */ |
| static int drm_count_cards(void) |
| { |
| int num = 0; |
| #if defined(DRIVER_CARD_LIST) |
| int i; |
| drm_pci_list_t *l; |
| u16 device, vendor; |
| struct pci_dev *pdev = NULL; |
| #endif |
| |
| DRM_DEBUG( "\n" ); |
| |
| #if defined(DRIVER_COUNT_CARDS) |
| num = DRIVER_COUNT_CARDS(); |
| #elif defined(DRIVER_CARD_LIST) |
| for (i = 0, l = DRIVER_CARD_LIST; l[i].vendor != 0; i++) { |
| pdev = NULL; |
| vendor = l[i].vendor; |
| device = l[i].device; |
| if(device == 0xffff) device = PCI_ANY_ID; |
| if(vendor == 0xffff) vendor = PCI_ANY_ID; |
| while ((pdev = pci_find_device(vendor, device, pdev))) { |
| num++; |
| } |
| } |
| #else |
| num = DRIVER_NUM_CARDS; |
| #endif |
| DRM_DEBUG("numdevs = %d\n", num); |
| return num; |
| } |
| |
| /* drm_init is called via init_module at module load time, or via |
| * linux/init/main.c (this is not currently supported). |
| */ |
| static int __init drm_init( void ) |
| { |
| |
| drm_device_t *dev; |
| int i; |
| #if __HAVE_CTX_BITMAP |
| int retcode; |
| #endif |
| DRM_DEBUG( "\n" ); |
| |
| #ifdef MODULE |
| DRM(parse_options)( drm_opts ); |
| #endif |
| |
| DRM(numdevs) = drm_count_cards(); |
| /* Force at least one instance. */ |
| if (DRM(numdevs) <= 0) |
| DRM(numdevs) = 1; |
| |
| DRM(device) = kmalloc(sizeof(*DRM(device)) * DRM(numdevs), GFP_KERNEL); |
| if (!DRM(device)) { |
| return -ENOMEM; |
| } |
| DRM(minor) = kmalloc(sizeof(*DRM(minor)) * DRM(numdevs), GFP_KERNEL); |
| if (!DRM(minor)) { |
| kfree(DRM(device)); |
| return -ENOMEM; |
| } |
| |
| DRIVER_PREINIT(); |
| |
| DRM(mem_init)(); |
| |
| for (i = 0; i < DRM(numdevs); i++) { |
| dev = &(DRM(device)[i]); |
| memset( (void *)dev, 0, sizeof(*dev) ); |
| dev->count_lock = SPIN_LOCK_UNLOCKED; |
| sema_init( &dev->struct_sem, 1 ); |
| |
| if ((DRM(minor)[i] = DRM(stub_register)(DRIVER_NAME, &DRM(fops),dev)) < 0) |
| return -EPERM; |
| dev->device = MKDEV(DRM_MAJOR, DRM(minor)[i] ); |
| dev->name = DRIVER_NAME; |
| |
| #if __REALLY_HAVE_AGP |
| dev->agp = DRM(agp_init)(); |
| #if __MUST_HAVE_AGP |
| if ( dev->agp == NULL ) { |
| DRM_ERROR( "Cannot initialize the agpgart module.\n" ); |
| DRM(stub_unregister)(DRM(minor)[i]); |
| DRM(takedown)( dev ); |
| return -ENOMEM; |
| } |
| #endif |
| #if __REALLY_HAVE_MTRR |
| if (dev->agp) |
| dev->agp->agp_mtrr = mtrr_add( dev->agp->agp_info.aper_base, |
| dev->agp->agp_info.aper_size*1024*1024, |
| MTRR_TYPE_WRCOMB, |
| 1 ); |
| #endif |
| #endif |
| |
| #if __HAVE_CTX_BITMAP |
| retcode = DRM(ctxbitmap_init)( dev ); |
| if( retcode ) { |
| DRM_ERROR( "Cannot allocate memory for context bitmap.\n" ); |
| DRM(stub_unregister)(DRM(minor)[i]); |
| DRM(takedown)( dev ); |
| return retcode; |
| } |
| #endif |
| DRM_INFO( "Initialized %s %d.%d.%d %s on minor %d\n", |
| DRIVER_NAME, |
| DRIVER_MAJOR, |
| DRIVER_MINOR, |
| DRIVER_PATCHLEVEL, |
| DRIVER_DATE, |
| DRM(minor)[i] ); |
| } |
| |
| DRIVER_POSTINIT(); |
| |
| return 0; |
| } |
| |
| /* drm_cleanup is called via cleanup_module at module unload time. |
| */ |
| static void __exit drm_cleanup( void ) |
| { |
| drm_device_t *dev; |
| int i; |
| |
| DRM_DEBUG( "\n" ); |
| |
| for (i = DRM(numdevs) - 1; i >= 0; i--) { |
| dev = &(DRM(device)[i]); |
| if ( DRM(stub_unregister)(DRM(minor)[i]) ) { |
| DRM_ERROR( "Cannot unload module\n" ); |
| } else { |
| DRM_DEBUG("minor %d unregistered\n", DRM(minor)[i]); |
| if (i == 0) { |
| DRM_INFO( "Module unloaded\n" ); |
| } |
| } |
| #if __HAVE_CTX_BITMAP |
| DRM(ctxbitmap_cleanup)( dev ); |
| #endif |
| |
| #if __REALLY_HAVE_AGP && __REALLY_HAVE_MTRR |
| if ( dev->agp && dev->agp->agp_mtrr >= 0) { |
| int retval; |
| retval = mtrr_del( dev->agp->agp_mtrr, |
| dev->agp->agp_info.aper_base, |
| dev->agp->agp_info.aper_size*1024*1024 ); |
| DRM_DEBUG( "mtrr_del=%d\n", retval ); |
| } |
| #endif |
| |
| DRM(takedown)( dev ); |
| |
| #if __REALLY_HAVE_AGP |
| if ( dev->agp ) { |
| DRM(agp_uninit)(); |
| DRM(free)( dev->agp, sizeof(*dev->agp), DRM_MEM_AGPLISTS ); |
| dev->agp = NULL; |
| } |
| #endif |
| } |
| DRIVER_POSTCLEANUP(); |
| kfree(DRM(minor)); |
| kfree(DRM(device)); |
| DRM(numdevs) = 0; |
| } |
| |
| module_init( drm_init ); |
| module_exit( drm_cleanup ); |
| |
| |
| int DRM(version)( struct inode *inode, struct file *filp, |
| unsigned int cmd, unsigned long arg ) |
| { |
| drm_version_t version; |
| int len; |
| |
| if ( copy_from_user( &version, |
| (drm_version_t *)arg, |
| sizeof(version) ) ) |
| return -EFAULT; |
| |
| #define DRM_COPY( name, value ) \ |
| len = strlen( value ); \ |
| if ( len > name##_len ) len = name##_len; \ |
| name##_len = strlen( value ); \ |
| if ( len && name ) { \ |
| if ( copy_to_user( name, value, len ) ) \ |
| return -EFAULT; \ |
| } |
| |
| version.version_major = DRIVER_MAJOR; |
| version.version_minor = DRIVER_MINOR; |
| version.version_patchlevel = DRIVER_PATCHLEVEL; |
| |
| DRM_COPY( version.name, DRIVER_NAME ); |
| DRM_COPY( version.date, DRIVER_DATE ); |
| DRM_COPY( version.desc, DRIVER_DESC ); |
| |
| if ( copy_to_user( (drm_version_t *)arg, |
| &version, |
| sizeof(version) ) ) |
| return -EFAULT; |
| return 0; |
| } |
| |
| int DRM(open)( struct inode *inode, struct file *filp ) |
| { |
| drm_device_t *dev = NULL; |
| int retcode = 0; |
| int i; |
| |
| for (i = 0; i < DRM(numdevs); i++) { |
| if (minor(inode->i_rdev) == DRM(minor)[i]) { |
| dev = &(DRM(device)[i]); |
| break; |
| } |
| } |
| if (!dev) { |
| return -ENODEV; |
| } |
| |
| DRM_DEBUG( "open_count = %d\n", dev->open_count ); |
| |
| retcode = DRM(open_helper)( inode, filp, dev ); |
| if ( !retcode ) { |
| atomic_inc( &dev->counts[_DRM_STAT_OPENS] ); |
| spin_lock( &dev->count_lock ); |
| if ( !dev->open_count++ ) { |
| spin_unlock( &dev->count_lock ); |
| return DRM(setup)( dev ); |
| } |
| spin_unlock( &dev->count_lock ); |
| } |
| |
| return retcode; |
| } |
| |
| int DRM(release)( struct inode *inode, struct file *filp ) |
| { |
| drm_file_t *priv = filp->private_data; |
| drm_device_t *dev; |
| int retcode = 0; |
| |
| lock_kernel(); |
| dev = priv->dev; |
| |
| DRM_DEBUG( "open_count = %d\n", dev->open_count ); |
| |
| DRIVER_PRERELEASE(); |
| |
| /* ======================================================== |
| * Begin inline drm_release |
| */ |
| |
| DRM_DEBUG( "pid = %d, device = 0x%x, open_count = %d\n", |
| current->pid, dev->device, dev->open_count ); |
| |
| if ( dev->lock.hw_lock && |
| _DRM_LOCK_IS_HELD(dev->lock.hw_lock->lock) && |
| dev->lock.pid == current->pid ) { |
| DRM_DEBUG( "Process %d dead, freeing lock for context %d\n", |
| current->pid, |
| _DRM_LOCKING_CONTEXT(dev->lock.hw_lock->lock) ); |
| #if __HAVE_RELEASE |
| DRIVER_RELEASE(); |
| #endif |
| DRM(lock_free)( dev, &dev->lock.hw_lock->lock, |
| _DRM_LOCKING_CONTEXT(dev->lock.hw_lock->lock) ); |
| |
| /* FIXME: may require heavy-handed reset of |
| hardware at this point, possibly |
| processed via a callback to the X |
| server. */ |
| } |
| #if __HAVE_RELEASE |
| else if ( dev->lock.hw_lock ) { |
| /* The lock is required to reclaim buffers */ |
| DECLARE_WAITQUEUE( entry, current ); |
| add_wait_queue( &dev->lock.lock_queue, &entry ); |
| for (;;) { |
| current->state = TASK_INTERRUPTIBLE; |
| if ( !dev->lock.hw_lock ) { |
| /* Device has been unregistered */ |
| retcode = -EINTR; |
| break; |
| } |
| if ( DRM(lock_take)( &dev->lock.hw_lock->lock, |
| DRM_KERNEL_CONTEXT ) ) { |
| dev->lock.pid = priv->pid; |
| dev->lock.lock_time = jiffies; |
| atomic_inc( &dev->counts[_DRM_STAT_LOCKS] ); |
| break; /* Got lock */ |
| } |
| /* Contention */ |
| #if 0 |
| atomic_inc( &dev->total_sleeps ); |
| #endif |
| schedule(); |
| if ( signal_pending( current ) ) { |
| retcode = -ERESTARTSYS; |
| break; |
| } |
| } |
| current->state = TASK_RUNNING; |
| remove_wait_queue( &dev->lock.lock_queue, &entry ); |
| if( !retcode ) { |
| DRIVER_RELEASE(); |
| DRM(lock_free)( dev, &dev->lock.hw_lock->lock, |
| DRM_KERNEL_CONTEXT ); |
| } |
| } |
| #elif __HAVE_DMA |
| DRM(reclaim_buffers)( dev, priv->pid ); |
| #endif |
| |
| DRM(fasync)( -1, filp, 0 ); |
| |
| down( &dev->struct_sem ); |
| if ( priv->remove_auth_on_close == 1 ) { |
| drm_file_t *temp = dev->file_first; |
| while ( temp ) { |
| temp->authenticated = 0; |
| temp = temp->next; |
| } |
| } |
| if ( priv->prev ) { |
| priv->prev->next = priv->next; |
| } else { |
| dev->file_first = priv->next; |
| } |
| if ( priv->next ) { |
| priv->next->prev = priv->prev; |
| } else { |
| dev->file_last = priv->prev; |
| } |
| up( &dev->struct_sem ); |
| |
| DRM(free)( priv, sizeof(*priv), DRM_MEM_FILES ); |
| |
| /* ======================================================== |
| * End inline drm_release |
| */ |
| |
| atomic_inc( &dev->counts[_DRM_STAT_CLOSES] ); |
| spin_lock( &dev->count_lock ); |
| if ( !--dev->open_count ) { |
| if ( atomic_read( &dev->ioctl_count ) || dev->blocked ) { |
| DRM_ERROR( "Device busy: %d %d\n", |
| atomic_read( &dev->ioctl_count ), |
| dev->blocked ); |
| spin_unlock( &dev->count_lock ); |
| unlock_kernel(); |
| return -EBUSY; |
| } |
| spin_unlock( &dev->count_lock ); |
| unlock_kernel(); |
| return DRM(takedown)( dev ); |
| } |
| spin_unlock( &dev->count_lock ); |
| |
| unlock_kernel(); |
| return retcode; |
| } |
| |
| /* DRM(ioctl) is called whenever a process performs an ioctl on /dev/drm. |
| */ |
| int DRM(ioctl)( struct inode *inode, struct file *filp, |
| unsigned int cmd, unsigned long arg ) |
| { |
| drm_file_t *priv = filp->private_data; |
| drm_device_t *dev = priv->dev; |
| drm_ioctl_desc_t *ioctl; |
| drm_ioctl_t *func; |
| int nr = DRM_IOCTL_NR(cmd); |
| int retcode = 0; |
| |
| atomic_inc( &dev->ioctl_count ); |
| atomic_inc( &dev->counts[_DRM_STAT_IOCTLS] ); |
| ++priv->ioctl_count; |
| |
| DRM_DEBUG( "pid=%d, cmd=0x%02x, nr=0x%02x, dev 0x%x, auth=%d\n", |
| current->pid, cmd, nr, dev->device, priv->authenticated ); |
| |
| if ( nr >= DRIVER_IOCTL_COUNT ) { |
| retcode = -EINVAL; |
| } else { |
| ioctl = &DRM(ioctls)[nr]; |
| func = ioctl->func; |
| |
| if ( !func ) { |
| DRM_DEBUG( "no function\n" ); |
| retcode = -EINVAL; |
| } else if ( ( ioctl->root_only && !capable( CAP_SYS_ADMIN ) )|| |
| ( ioctl->auth_needed && !priv->authenticated ) ) { |
| retcode = -EACCES; |
| } else { |
| retcode = func( inode, filp, cmd, arg ); |
| } |
| } |
| |
| atomic_dec( &dev->ioctl_count ); |
| return retcode; |
| } |
| |
| int DRM(lock)( struct inode *inode, struct file *filp, |
| unsigned int cmd, unsigned long arg ) |
| { |
| drm_file_t *priv = filp->private_data; |
| drm_device_t *dev = priv->dev; |
| DECLARE_WAITQUEUE( entry, current ); |
| drm_lock_t lock; |
| int ret = 0; |
| #if __HAVE_MULTIPLE_DMA_QUEUES |
| drm_queue_t *q; |
| #endif |
| #if __HAVE_DMA_HISTOGRAM |
| cycles_t start; |
| |
| dev->lck_start = start = get_cycles(); |
| #endif |
| |
| if ( copy_from_user( &lock, (drm_lock_t *)arg, sizeof(lock) ) ) |
| return -EFAULT; |
| |
| if ( lock.context == DRM_KERNEL_CONTEXT ) { |
| DRM_ERROR( "Process %d using kernel context %d\n", |
| current->pid, lock.context ); |
| return -EINVAL; |
| } |
| |
| DRM_DEBUG( "%d (pid %d) requests lock (0x%08x), flags = 0x%08x\n", |
| lock.context, current->pid, |
| dev->lock.hw_lock->lock, lock.flags ); |
| |
| #if __HAVE_DMA_QUEUE |
| if ( lock.context < 0 ) |
| return -EINVAL; |
| #elif __HAVE_MULTIPLE_DMA_QUEUES |
| if ( lock.context < 0 || lock.context >= dev->queue_count ) |
| return -EINVAL; |
| q = dev->queuelist[lock.context]; |
| #endif |
| |
| #if __HAVE_DMA_FLUSH |
| ret = DRM(flush_block_and_flush)( dev, lock.context, lock.flags ); |
| #endif |
| if ( !ret ) { |
| add_wait_queue( &dev->lock.lock_queue, &entry ); |
| for (;;) { |
| current->state = TASK_INTERRUPTIBLE; |
| if ( !dev->lock.hw_lock ) { |
| /* Device has been unregistered */ |
| ret = -EINTR; |
| break; |
| } |
| if ( DRM(lock_take)( &dev->lock.hw_lock->lock, |
| lock.context ) ) { |
| dev->lock.pid = current->pid; |
| dev->lock.lock_time = jiffies; |
| atomic_inc( &dev->counts[_DRM_STAT_LOCKS] ); |
| break; /* Got lock */ |
| } |
| |
| /* Contention */ |
| schedule(); |
| if ( signal_pending( current ) ) { |
| ret = -ERESTARTSYS; |
| break; |
| } |
| } |
| current->state = TASK_RUNNING; |
| remove_wait_queue( &dev->lock.lock_queue, &entry ); |
| } |
| |
| #if __HAVE_DMA_FLUSH |
| DRM(flush_unblock)( dev, lock.context, lock.flags ); /* cleanup phase */ |
| #endif |
| |
| if ( !ret ) { |
| sigemptyset( &dev->sigmask ); |
| sigaddset( &dev->sigmask, SIGSTOP ); |
| sigaddset( &dev->sigmask, SIGTSTP ); |
| sigaddset( &dev->sigmask, SIGTTIN ); |
| sigaddset( &dev->sigmask, SIGTTOU ); |
| dev->sigdata.context = lock.context; |
| dev->sigdata.lock = dev->lock.hw_lock; |
| block_all_signals( DRM(notifier), |
| &dev->sigdata, &dev->sigmask ); |
| |
| #if __HAVE_DMA_READY |
| if ( lock.flags & _DRM_LOCK_READY ) { |
| DRIVER_DMA_READY(); |
| } |
| #endif |
| #if __HAVE_DMA_QUIESCENT |
| if ( lock.flags & _DRM_LOCK_QUIESCENT ) { |
| DRIVER_DMA_QUIESCENT(); |
| } |
| #endif |
| #if __HAVE_KERNEL_CTX_SWITCH |
| if ( dev->last_context != lock.context ) { |
| DRM(context_switch)(dev, dev->last_context, |
| lock.context); |
| } |
| #endif |
| } |
| |
| DRM_DEBUG( "%d %s\n", lock.context, ret ? "interrupted" : "has lock" ); |
| |
| #if __HAVE_DMA_HISTOGRAM |
| atomic_inc(&dev->histo.lacq[DRM(histogram_slot)(get_cycles()-start)]); |
| #endif |
| return ret; |
| } |
| |
| |
| int DRM(unlock)( struct inode *inode, struct file *filp, |
| unsigned int cmd, unsigned long arg ) |
| { |
| drm_file_t *priv = filp->private_data; |
| drm_device_t *dev = priv->dev; |
| drm_lock_t lock; |
| |
| if ( copy_from_user( &lock, (drm_lock_t *)arg, sizeof(lock) ) ) |
| return -EFAULT; |
| |
| if ( lock.context == DRM_KERNEL_CONTEXT ) { |
| DRM_ERROR( "Process %d using kernel context %d\n", |
| current->pid, lock.context ); |
| return -EINVAL; |
| } |
| |
| atomic_inc( &dev->counts[_DRM_STAT_UNLOCKS] ); |
| |
| #if __HAVE_KERNEL_CTX_SWITCH |
| /* We no longer really hold it, but if we are the next |
| * agent to request it then we should just be able to |
| * take it immediately and not eat the ioctl. |
| */ |
| dev->lock.pid = 0; |
| { |
| __volatile__ unsigned int *plock = &dev->lock.hw_lock->lock; |
| unsigned int old, new, prev, ctx; |
| |
| ctx = lock.context; |
| do { |
| old = *plock; |
| new = ctx; |
| prev = cmpxchg(plock, old, new); |
| } while (prev != old); |
| } |
| wake_up_interruptible(&dev->lock.lock_queue); |
| #else |
| DRM(lock_transfer)( dev, &dev->lock.hw_lock->lock, |
| DRM_KERNEL_CONTEXT ); |
| #if __HAVE_DMA_SCHEDULE |
| DRM(dma_schedule)( dev, 1 ); |
| #endif |
| |
| /* FIXME: Do we ever really need to check this??? |
| */ |
| if ( 1 /* !dev->context_flag */ ) { |
| if ( DRM(lock_free)( dev, &dev->lock.hw_lock->lock, |
| DRM_KERNEL_CONTEXT ) ) { |
| DRM_ERROR( "\n" ); |
| } |
| } |
| #endif /* !__HAVE_KERNEL_CTX_SWITCH */ |
| |
| unblock_all_signals(); |
| return 0; |
| } |