| /********************************************************** |
| * Copyright 2008-2009 VMware, Inc. All rights reserved. |
| * |
| * Permission is hereby granted, free of charge, to any person |
| * obtaining a copy of this software and associated documentation |
| * files (the "Software"), to deal in the Software without |
| * restriction, including without limitation the rights to use, copy, |
| * modify, merge, publish, distribute, sublicense, and/or sell copies |
| * of the Software, and to permit persons to whom the Software is |
| * furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be |
| * included in all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
| * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND |
| * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS |
| * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN |
| * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
| * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| * |
| **********************************************************/ |
| |
| #include "svga_cmd.h" |
| |
| #include "pipe/p_state.h" |
| #include "pipe/p_defines.h" |
| #include "pipe/p_inlines.h" |
| #include "pipe/p_thread.h" |
| #include "util/u_math.h" |
| #include "util/u_memory.h" |
| |
| #include "svga_context.h" |
| #include "svga_screen.h" |
| #include "svga_screen_buffer.h" |
| #include "svga_winsys.h" |
| #include "svga_debug.h" |
| |
| |
| /** |
| * Vertex and index buffers have to be treated slightly differently from |
| * regular guest memory regions because the SVGA device sees them as |
| * surfaces, and the state tracker can create/destroy without the pipe |
| * driver, therefore we must do the uploads from the vws. |
| */ |
| static INLINE boolean |
| svga_buffer_needs_hw_storage(unsigned usage) |
| { |
| return usage & (PIPE_BUFFER_USAGE_VERTEX | PIPE_BUFFER_USAGE_INDEX); |
| } |
| |
| |
| static INLINE enum pipe_error |
| svga_buffer_create_host_surface(struct svga_screen *ss, |
| struct svga_buffer *sbuf) |
| { |
| if(!sbuf->handle) { |
| sbuf->key.flags = 0; |
| |
| sbuf->key.format = SVGA3D_BUFFER; |
| if(sbuf->base.usage & PIPE_BUFFER_USAGE_VERTEX) |
| sbuf->key.flags |= SVGA3D_SURFACE_HINT_VERTEXBUFFER; |
| if(sbuf->base.usage & PIPE_BUFFER_USAGE_INDEX) |
| sbuf->key.flags |= SVGA3D_SURFACE_HINT_INDEXBUFFER; |
| |
| sbuf->key.size.width = sbuf->base.size; |
| sbuf->key.size.height = 1; |
| sbuf->key.size.depth = 1; |
| |
| sbuf->key.numFaces = 1; |
| sbuf->key.numMipLevels = 1; |
| sbuf->key.cachable = 1; |
| |
| SVGA_DBG(DEBUG_DMA, "surface_create for buffer sz %d\n", sbuf->base.size); |
| |
| sbuf->handle = svga_screen_surface_create(ss, &sbuf->key); |
| if(!sbuf->handle) |
| return PIPE_ERROR_OUT_OF_MEMORY; |
| |
| /* Always set the discard flag on the first time the buffer is written |
| * as svga_screen_surface_create might have passed a recycled host |
| * buffer. |
| */ |
| sbuf->hw.flags.discard = TRUE; |
| |
| SVGA_DBG(DEBUG_DMA, " --> got sid %p sz %d (buffer)\n", sbuf->handle, sbuf->base.size); |
| } |
| |
| return PIPE_OK; |
| } |
| |
| |
| static INLINE void |
| svga_buffer_destroy_host_surface(struct svga_screen *ss, |
| struct svga_buffer *sbuf) |
| { |
| if(sbuf->handle) { |
| SVGA_DBG(DEBUG_DMA, " ungrab sid %p sz %d\n", sbuf->handle, sbuf->base.size); |
| svga_screen_surface_destroy(ss, &sbuf->key, &sbuf->handle); |
| } |
| } |
| |
| |
| static INLINE void |
| svga_buffer_destroy_hw_storage(struct svga_screen *ss, struct svga_buffer *sbuf) |
| { |
| struct svga_winsys_screen *sws = ss->sws; |
| |
| assert(!sbuf->map.count); |
| assert(sbuf->hw.buf); |
| if(sbuf->hw.buf) { |
| sws->buffer_destroy(sws, sbuf->hw.buf); |
| sbuf->hw.buf = NULL; |
| assert(sbuf->head.prev && sbuf->head.next); |
| LIST_DEL(&sbuf->head); |
| #ifdef DEBUG |
| sbuf->head.next = sbuf->head.prev = NULL; |
| #endif |
| } |
| } |
| |
| static INLINE enum pipe_error |
| svga_buffer_backup(struct svga_screen *ss, struct svga_buffer *sbuf) |
| { |
| if (sbuf->hw.buf && sbuf->hw.num_ranges) { |
| void *src; |
| |
| if (!sbuf->swbuf) |
| sbuf->swbuf = align_malloc(sbuf->base.size, sbuf->base.alignment); |
| if (!sbuf->swbuf) |
| return PIPE_ERROR_OUT_OF_MEMORY; |
| |
| src = ss->sws->buffer_map(ss->sws, sbuf->hw.buf, |
| PIPE_BUFFER_USAGE_CPU_READ); |
| if (!src) |
| return PIPE_ERROR; |
| |
| memcpy(sbuf->swbuf, src, sbuf->base.size); |
| ss->sws->buffer_unmap(ss->sws, sbuf->hw.buf); |
| } |
| |
| return PIPE_OK; |
| } |
| |
| /** |
| * Try to make GMR space available by freeing the hardware storage of |
| * unmapped |
| */ |
| boolean |
| svga_buffer_free_cached_hw_storage(struct svga_screen *ss) |
| { |
| struct list_head *curr; |
| struct svga_buffer *sbuf; |
| enum pipe_error ret = PIPE_OK; |
| |
| curr = ss->cached_buffers.prev; |
| |
| /* free the least recently used buffer's hw storage which is not mapped */ |
| do { |
| if(curr == &ss->cached_buffers) |
| return FALSE; |
| |
| sbuf = LIST_ENTRY(struct svga_buffer, curr, head); |
| |
| curr = curr->prev; |
| if (sbuf->map.count == 0) |
| ret = svga_buffer_backup(ss, sbuf); |
| |
| } while(sbuf->map.count != 0 || ret != PIPE_OK); |
| |
| svga_buffer_destroy_hw_storage(ss, sbuf); |
| |
| return TRUE; |
| } |
| |
| struct svga_winsys_buffer * |
| svga_winsys_buffer_create( struct svga_screen *ss, |
| unsigned alignment, |
| unsigned usage, |
| unsigned size ) |
| { |
| struct svga_winsys_screen *sws = ss->sws; |
| struct svga_winsys_buffer *buf; |
| |
| /* Just try */ |
| buf = sws->buffer_create(sws, alignment, usage, size); |
| if(!buf) { |
| |
| SVGA_DBG(DEBUG_DMA|DEBUG_PERF, "flushing screen to find %d bytes GMR\n", |
| size); |
| |
| /* Try flushing all pending DMAs */ |
| svga_screen_flush(ss, NULL); |
| buf = sws->buffer_create(sws, alignment, usage, size); |
| |
| SVGA_DBG(DEBUG_DMA|DEBUG_PERF, "evicting buffers to find %d bytes GMR\n", |
| size); |
| |
| /* Try evicing all buffer storage */ |
| while(!buf && svga_buffer_free_cached_hw_storage(ss)) |
| buf = sws->buffer_create(sws, alignment, usage, size); |
| } |
| |
| return buf; |
| } |
| |
| |
| /** |
| * Allocate DMA'ble storage for the buffer. |
| * |
| * Called before mapping a buffer. |
| */ |
| static INLINE enum pipe_error |
| svga_buffer_create_hw_storage(struct svga_screen *ss, |
| struct svga_buffer *sbuf) |
| { |
| if(!sbuf->hw.buf) { |
| unsigned alignment = sbuf->base.alignment; |
| unsigned usage = 0; |
| unsigned size = sbuf->base.size; |
| |
| sbuf->hw.buf = svga_winsys_buffer_create(ss, alignment, usage, size); |
| if(!sbuf->hw.buf) |
| return PIPE_ERROR_OUT_OF_MEMORY; |
| |
| assert(!sbuf->needs_flush); |
| assert(!sbuf->head.prev && !sbuf->head.next); |
| LIST_ADD(&sbuf->head, &ss->cached_buffers); |
| } |
| |
| return PIPE_OK; |
| } |
| |
| |
| /** |
| * Variant of SVGA3D_BufferDMA which leaves the copy box temporarily in blank. |
| */ |
| static enum pipe_error |
| svga_buffer_upload_command(struct svga_context *svga, |
| struct svga_buffer *sbuf) |
| { |
| struct svga_winsys_context *swc = svga->swc; |
| struct svga_winsys_buffer *guest = sbuf->hw.buf; |
| struct svga_winsys_surface *host = sbuf->handle; |
| SVGA3dTransferType transfer = SVGA3D_WRITE_HOST_VRAM; |
| SVGA3dSurfaceDMAFlags flags = sbuf->hw.flags; |
| SVGA3dCmdSurfaceDMA *cmd; |
| uint32 numBoxes = sbuf->hw.num_ranges; |
| SVGA3dCopyBox *boxes; |
| SVGA3dCmdSurfaceDMASuffix *pSuffix; |
| unsigned region_flags; |
| unsigned surface_flags; |
| struct pipe_buffer *dummy; |
| |
| if(transfer == SVGA3D_WRITE_HOST_VRAM) { |
| region_flags = PIPE_BUFFER_USAGE_GPU_READ; |
| surface_flags = PIPE_BUFFER_USAGE_GPU_WRITE; |
| } |
| else if(transfer == SVGA3D_READ_HOST_VRAM) { |
| region_flags = PIPE_BUFFER_USAGE_GPU_WRITE; |
| surface_flags = PIPE_BUFFER_USAGE_GPU_READ; |
| } |
| else { |
| assert(0); |
| return PIPE_ERROR_BAD_INPUT; |
| } |
| |
| assert(numBoxes); |
| |
| cmd = SVGA3D_FIFOReserve(swc, |
| SVGA_3D_CMD_SURFACE_DMA, |
| sizeof *cmd + numBoxes * sizeof *boxes + sizeof *pSuffix, |
| 2); |
| if(!cmd) |
| return PIPE_ERROR_OUT_OF_MEMORY; |
| |
| swc->region_relocation(swc, &cmd->guest.ptr, guest, 0, region_flags); |
| cmd->guest.pitch = 0; |
| |
| swc->surface_relocation(swc, &cmd->host.sid, host, surface_flags); |
| cmd->host.face = 0; |
| cmd->host.mipmap = 0; |
| |
| cmd->transfer = transfer; |
| |
| sbuf->hw.boxes = (SVGA3dCopyBox *)&cmd[1]; |
| sbuf->hw.svga = svga; |
| |
| /* Increment reference count */ |
| dummy = NULL; |
| pipe_buffer_reference(&dummy, &sbuf->base); |
| |
| pSuffix = (SVGA3dCmdSurfaceDMASuffix *)((uint8_t*)cmd + sizeof *cmd + numBoxes * sizeof *boxes); |
| pSuffix->suffixSize = sizeof *pSuffix; |
| pSuffix->maximumOffset = sbuf->base.size; |
| pSuffix->flags = flags; |
| |
| swc->commit(swc); |
| |
| return PIPE_OK; |
| } |
| |
| |
| /** |
| * Patch up the upload DMA command reserved by svga_buffer_upload_command |
| * with the final ranges. |
| */ |
| static void |
| svga_buffer_upload_flush(struct svga_context *svga, |
| struct svga_buffer *sbuf) |
| { |
| struct svga_screen *ss = svga_screen(svga->pipe.screen); |
| SVGA3dCopyBox *boxes; |
| unsigned i; |
| |
| assert(sbuf->handle); |
| assert(sbuf->hw.buf); |
| assert(sbuf->hw.num_ranges); |
| assert(sbuf->hw.svga == svga); |
| assert(sbuf->hw.boxes); |
| |
| /* |
| * Patch the DMA command with the final copy box. |
| */ |
| |
| SVGA_DBG(DEBUG_DMA, "dma to sid %p\n", sbuf->handle); |
| |
| boxes = sbuf->hw.boxes; |
| for(i = 0; i < sbuf->hw.num_ranges; ++i) { |
| SVGA_DBG(DEBUG_DMA, " bytes %u - %u\n", |
| sbuf->hw.ranges[i].start, sbuf->hw.ranges[i].end); |
| |
| boxes[i].x = sbuf->hw.ranges[i].start; |
| boxes[i].y = 0; |
| boxes[i].z = 0; |
| boxes[i].w = sbuf->hw.ranges[i].end - sbuf->hw.ranges[i].start; |
| boxes[i].h = 1; |
| boxes[i].d = 1; |
| boxes[i].srcx = sbuf->hw.ranges[i].start; |
| boxes[i].srcy = 0; |
| boxes[i].srcz = 0; |
| } |
| |
| sbuf->hw.num_ranges = 0; |
| memset(&sbuf->hw.flags, 0, sizeof sbuf->hw.flags); |
| |
| assert(sbuf->head.prev && sbuf->head.next); |
| LIST_DEL(&sbuf->head); |
| sbuf->needs_flush = FALSE; |
| /* XXX: do we care about cached_buffers any more ?*/ |
| LIST_ADD(&sbuf->head, &ss->cached_buffers); |
| |
| sbuf->hw.svga = NULL; |
| sbuf->hw.boxes = NULL; |
| |
| /* Decrement reference count */ |
| pipe_reference(&(sbuf->base.reference), NULL); |
| sbuf = NULL; |
| } |
| |
| |
| /** |
| * Queue a DMA upload of a range of this buffer to the host. |
| * |
| * This function only notes the range down. It doesn't actually emit a DMA |
| * upload command. That only happens when a context tries to refer to this |
| * buffer, and the DMA upload command is added to that context's command buffer. |
| * |
| * We try to lump as many contiguous DMA transfers together as possible. |
| */ |
| static void |
| svga_buffer_upload_queue(struct svga_buffer *sbuf, |
| unsigned start, |
| unsigned end) |
| { |
| unsigned i; |
| |
| assert(sbuf->hw.buf); |
| assert(end > start); |
| |
| /* |
| * Try to grow one of the ranges. |
| * |
| * Note that it is not this function task to care about overlapping ranges, |
| * as the GMR was already given so it is too late to do anything. Situations |
| * where overlapping ranges may pose a problem should be detected via |
| * pipe_context::is_buffer_referenced and the context that refers to the |
| * buffer should be flushed. |
| */ |
| |
| for(i = 0; i < sbuf->hw.num_ranges; ++i) { |
| if(start <= sbuf->hw.ranges[i].end && sbuf->hw.ranges[i].start <= end) { |
| sbuf->hw.ranges[i].start = MIN2(sbuf->hw.ranges[i].start, start); |
| sbuf->hw.ranges[i].end = MAX2(sbuf->hw.ranges[i].end, end); |
| return; |
| } |
| } |
| |
| /* |
| * We cannot add a new range to an existing DMA command, so patch-up the |
| * pending DMA upload and start clean. |
| */ |
| |
| if(sbuf->needs_flush) |
| svga_buffer_upload_flush(sbuf->hw.svga, sbuf); |
| |
| assert(!sbuf->needs_flush); |
| assert(!sbuf->hw.svga); |
| assert(!sbuf->hw.boxes); |
| |
| /* |
| * Add a new range. |
| */ |
| |
| sbuf->hw.ranges[sbuf->hw.num_ranges].start = start; |
| sbuf->hw.ranges[sbuf->hw.num_ranges].end = end; |
| ++sbuf->hw.num_ranges; |
| } |
| |
| |
| static void * |
| svga_buffer_map_range( struct pipe_screen *screen, |
| struct pipe_buffer *buf, |
| unsigned offset, unsigned length, |
| unsigned usage ) |
| { |
| struct svga_screen *ss = svga_screen(screen); |
| struct svga_winsys_screen *sws = ss->sws; |
| struct svga_buffer *sbuf = svga_buffer( buf ); |
| void *map; |
| |
| if(sbuf->swbuf) { |
| /* User/malloc buffer */ |
| map = sbuf->swbuf; |
| } |
| else { |
| if(!sbuf->hw.buf) { |
| struct svga_winsys_surface *handle = sbuf->handle; |
| |
| if(svga_buffer_create_hw_storage(ss, sbuf) != PIPE_OK) |
| return NULL; |
| |
| /* Populate the hardware storage if the host surface pre-existed */ |
| if((usage & PIPE_BUFFER_USAGE_CPU_READ) && handle) { |
| SVGA3dSurfaceDMAFlags flags; |
| enum pipe_error ret; |
| struct pipe_fence_handle *fence = NULL; |
| |
| SVGA_DBG(DEBUG_DMA|DEBUG_PERF, "dma from sid %p (buffer), bytes %u - %u\n", |
| sbuf->handle, 0, sbuf->base.size); |
| |
| memset(&flags, 0, sizeof flags); |
| |
| ret = SVGA3D_BufferDMA(ss->swc, |
| sbuf->hw.buf, |
| sbuf->handle, |
| SVGA3D_READ_HOST_VRAM, |
| sbuf->base.size, |
| 0, |
| flags); |
| if(ret != PIPE_OK) { |
| ss->swc->flush(ss->swc, NULL); |
| |
| ret = SVGA3D_BufferDMA(ss->swc, |
| sbuf->hw.buf, |
| sbuf->handle, |
| SVGA3D_READ_HOST_VRAM, |
| sbuf->base.size, |
| 0, |
| flags); |
| assert(ret == PIPE_OK); |
| } |
| |
| ss->swc->flush(ss->swc, &fence); |
| sws->fence_finish(sws, fence, 0); |
| sws->fence_reference(sws, &fence, NULL); |
| } |
| } |
| else { |
| if((usage & PIPE_BUFFER_USAGE_CPU_READ) && !sbuf->needs_flush) { |
| /* We already had the hardware storage but we would have to issue |
| * a download if we hadn't, so move the buffer to the begginning |
| * of the LRU list. |
| */ |
| assert(sbuf->head.prev && sbuf->head.next); |
| LIST_DEL(&sbuf->head); |
| LIST_ADD(&sbuf->head, &ss->cached_buffers); |
| } |
| } |
| |
| map = sws->buffer_map(sws, sbuf->hw.buf, usage); |
| } |
| |
| if(map) { |
| pipe_mutex_lock(ss->swc_mutex); |
| |
| ++sbuf->map.count; |
| |
| if (usage & PIPE_BUFFER_USAGE_CPU_WRITE) { |
| assert(sbuf->map.count <= 1); |
| sbuf->map.writing = TRUE; |
| if (usage & PIPE_BUFFER_USAGE_FLUSH_EXPLICIT) |
| sbuf->map.flush_explicit = TRUE; |
| } |
| |
| pipe_mutex_unlock(ss->swc_mutex); |
| } |
| |
| return map; |
| } |
| |
| static void |
| svga_buffer_flush_mapped_range( struct pipe_screen *screen, |
| struct pipe_buffer *buf, |
| unsigned offset, unsigned length) |
| { |
| struct svga_buffer *sbuf = svga_buffer( buf ); |
| struct svga_screen *ss = svga_screen(screen); |
| |
| pipe_mutex_lock(ss->swc_mutex); |
| assert(sbuf->map.writing); |
| if(sbuf->map.writing) { |
| assert(sbuf->map.flush_explicit); |
| if(sbuf->hw.buf) |
| svga_buffer_upload_queue(sbuf, offset, offset + length); |
| } |
| pipe_mutex_unlock(ss->swc_mutex); |
| } |
| |
| static void |
| svga_buffer_unmap( struct pipe_screen *screen, |
| struct pipe_buffer *buf) |
| { |
| struct svga_screen *ss = svga_screen(screen); |
| struct svga_winsys_screen *sws = ss->sws; |
| struct svga_buffer *sbuf = svga_buffer( buf ); |
| |
| pipe_mutex_lock(ss->swc_mutex); |
| |
| assert(sbuf->map.count); |
| if(sbuf->map.count) |
| --sbuf->map.count; |
| |
| if(sbuf->hw.buf) |
| sws->buffer_unmap(sws, sbuf->hw.buf); |
| |
| if(sbuf->map.writing) { |
| if(!sbuf->map.flush_explicit) { |
| /* No mapped range was flushed -- flush the whole buffer */ |
| SVGA_DBG(DEBUG_DMA, "flushing the whole buffer\n"); |
| |
| if(sbuf->hw.buf) |
| svga_buffer_upload_queue(sbuf, 0, sbuf->base.size); |
| } |
| |
| sbuf->map.writing = FALSE; |
| sbuf->map.flush_explicit = FALSE; |
| } |
| |
| pipe_mutex_unlock(ss->swc_mutex); |
| } |
| |
| static void |
| svga_buffer_destroy( struct pipe_buffer *buf ) |
| { |
| struct svga_screen *ss = svga_screen(buf->screen); |
| struct svga_buffer *sbuf = svga_buffer( buf ); |
| |
| assert(!p_atomic_read(&buf->reference.count)); |
| |
| assert(!sbuf->needs_flush); |
| |
| if(sbuf->handle) { |
| SVGA_DBG(DEBUG_DMA, "release sid %p sz %d\n", sbuf->handle, sbuf->base.size); |
| svga_screen_surface_destroy(ss, &sbuf->key, &sbuf->handle); |
| } |
| |
| if(sbuf->hw.buf) |
| svga_buffer_destroy_hw_storage(ss, sbuf); |
| |
| if(sbuf->swbuf && !sbuf->user) |
| align_free(sbuf->swbuf); |
| |
| FREE(sbuf); |
| } |
| |
| static struct pipe_buffer * |
| svga_buffer_create(struct pipe_screen *screen, |
| unsigned alignment, |
| unsigned usage, |
| unsigned size) |
| { |
| struct svga_screen *ss = svga_screen(screen); |
| struct svga_buffer *sbuf; |
| |
| sbuf = CALLOC_STRUCT(svga_buffer); |
| if(!sbuf) |
| goto error1; |
| |
| sbuf->magic = SVGA_BUFFER_MAGIC; |
| |
| pipe_reference_init(&sbuf->base.reference, 1); |
| sbuf->base.screen = screen; |
| sbuf->base.alignment = alignment; |
| sbuf->base.usage = usage; |
| sbuf->base.size = size; |
| |
| if(svga_buffer_needs_hw_storage(usage)) { |
| if(svga_buffer_create_host_surface(ss, sbuf) != PIPE_OK) |
| goto error2; |
| } |
| else { |
| if(alignment < sizeof(void*)) |
| alignment = sizeof(void*); |
| |
| usage |= PIPE_BUFFER_USAGE_CPU_READ_WRITE; |
| |
| sbuf->swbuf = align_malloc(size, alignment); |
| if(!sbuf->swbuf) |
| goto error2; |
| } |
| |
| return &sbuf->base; |
| |
| error2: |
| FREE(sbuf); |
| error1: |
| return NULL; |
| } |
| |
| static struct pipe_buffer * |
| svga_user_buffer_create(struct pipe_screen *screen, |
| void *ptr, |
| unsigned bytes) |
| { |
| struct svga_buffer *sbuf; |
| |
| sbuf = CALLOC_STRUCT(svga_buffer); |
| if(!sbuf) |
| goto no_sbuf; |
| |
| sbuf->magic = SVGA_BUFFER_MAGIC; |
| |
| sbuf->swbuf = ptr; |
| sbuf->user = TRUE; |
| |
| pipe_reference_init(&sbuf->base.reference, 1); |
| sbuf->base.screen = screen; |
| sbuf->base.alignment = 1; |
| sbuf->base.usage = 0; |
| sbuf->base.size = bytes; |
| |
| return &sbuf->base; |
| |
| no_sbuf: |
| return NULL; |
| } |
| |
| |
| void |
| svga_screen_init_buffer_functions(struct pipe_screen *screen) |
| { |
| screen->buffer_create = svga_buffer_create; |
| screen->user_buffer_create = svga_user_buffer_create; |
| screen->buffer_map_range = svga_buffer_map_range; |
| screen->buffer_flush_mapped_range = svga_buffer_flush_mapped_range; |
| screen->buffer_unmap = svga_buffer_unmap; |
| screen->buffer_destroy = svga_buffer_destroy; |
| } |
| |
| |
| /** |
| * Copy the contents of the user buffer / malloc buffer to a hardware buffer. |
| */ |
| static INLINE enum pipe_error |
| svga_buffer_update_hw(struct svga_screen *ss, struct svga_buffer *sbuf) |
| { |
| if(!sbuf->hw.buf) { |
| enum pipe_error ret; |
| void *map; |
| |
| assert(sbuf->swbuf); |
| if(!sbuf->swbuf) |
| return PIPE_ERROR; |
| |
| ret = svga_buffer_create_hw_storage(ss, sbuf); |
| assert(ret == PIPE_OK); |
| if(ret != PIPE_OK) |
| return ret; |
| |
| pipe_mutex_lock(ss->swc_mutex); |
| map = ss->sws->buffer_map(ss->sws, sbuf->hw.buf, PIPE_BUFFER_USAGE_CPU_WRITE); |
| assert(map); |
| if(!map) { |
| pipe_mutex_unlock(ss->swc_mutex); |
| return PIPE_ERROR_OUT_OF_MEMORY; |
| } |
| |
| memcpy(map, sbuf->swbuf, sbuf->base.size); |
| ss->sws->buffer_unmap(ss->sws, sbuf->hw.buf); |
| |
| /* This user/malloc buffer is now indistinguishable from a gpu buffer */ |
| assert(!sbuf->map.count); |
| if(!sbuf->map.count) { |
| if(sbuf->user) |
| sbuf->user = FALSE; |
| else |
| align_free(sbuf->swbuf); |
| sbuf->swbuf = NULL; |
| } |
| |
| svga_buffer_upload_queue(sbuf, 0, sbuf->base.size); |
| } |
| |
| pipe_mutex_unlock(ss->swc_mutex); |
| return PIPE_OK; |
| } |
| |
| |
| struct svga_winsys_surface * |
| svga_buffer_handle(struct svga_context *svga, |
| struct pipe_buffer *buf) |
| { |
| struct pipe_screen *screen = svga->pipe.screen; |
| struct svga_screen *ss = svga_screen(screen); |
| struct svga_buffer *sbuf; |
| enum pipe_error ret; |
| |
| if(!buf) |
| return NULL; |
| |
| sbuf = svga_buffer(buf); |
| |
| assert(!sbuf->map.count); |
| |
| if(!sbuf->handle) { |
| ret = svga_buffer_create_host_surface(ss, sbuf); |
| if(ret != PIPE_OK) |
| return NULL; |
| |
| ret = svga_buffer_update_hw(ss, sbuf); |
| if(ret != PIPE_OK) |
| return NULL; |
| } |
| |
| if(!sbuf->needs_flush && sbuf->hw.num_ranges) { |
| /* Queue the buffer for flushing */ |
| ret = svga_buffer_upload_command(svga, sbuf); |
| if(ret != PIPE_OK) |
| /* XXX: Should probably have a richer return value */ |
| return NULL; |
| |
| assert(sbuf->hw.svga == svga); |
| |
| sbuf->needs_flush = TRUE; |
| assert(sbuf->head.prev && sbuf->head.next); |
| LIST_DEL(&sbuf->head); |
| LIST_ADDTAIL(&sbuf->head, &svga->dirty_buffers); |
| } |
| |
| return sbuf->handle; |
| } |
| |
| struct pipe_buffer * |
| svga_screen_buffer_wrap_surface(struct pipe_screen *screen, |
| enum SVGA3dSurfaceFormat format, |
| struct svga_winsys_surface *srf) |
| { |
| struct pipe_buffer *buf; |
| struct svga_buffer *sbuf; |
| struct svga_winsys_screen *sws = svga_winsys_screen(screen); |
| |
| buf = svga_buffer_create(screen, 0, SVGA_BUFFER_USAGE_WRAPPED, 0); |
| if (!buf) |
| return NULL; |
| |
| sbuf = svga_buffer(buf); |
| |
| /* |
| * We are not the creator of this surface and therefore we must not |
| * cache it for reuse. Set the cacheable flag to zero in the key to |
| * prevent this. |
| */ |
| sbuf->key.format = format; |
| sbuf->key.cachable = 0; |
| sws->surface_reference(sws, &sbuf->handle, srf); |
| |
| return buf; |
| } |
| |
| |
| struct svga_winsys_surface * |
| svga_screen_buffer_get_winsys_surface(struct pipe_buffer *buffer) |
| { |
| struct svga_winsys_screen *sws = svga_winsys_screen(buffer->screen); |
| struct svga_winsys_surface *vsurf = NULL; |
| |
| assert(svga_buffer(buffer)->key.cachable == 0); |
| svga_buffer(buffer)->key.cachable = 0; |
| sws->surface_reference(sws, &vsurf, svga_buffer(buffer)->handle); |
| return vsurf; |
| } |
| |
| void |
| svga_context_flush_buffers(struct svga_context *svga) |
| { |
| struct list_head *curr, *next; |
| struct svga_buffer *sbuf; |
| |
| curr = svga->dirty_buffers.next; |
| next = curr->next; |
| while(curr != &svga->dirty_buffers) { |
| sbuf = LIST_ENTRY(struct svga_buffer, curr, head); |
| |
| assert(p_atomic_read(&sbuf->base.reference.count) != 0); |
| assert(sbuf->needs_flush); |
| |
| svga_buffer_upload_flush(svga, sbuf); |
| |
| curr = next; |
| next = curr->next; |
| } |
| } |