struct pipe_transfer * r600_compute_global_get_transfer( struct pipe_context *ctx_, struct pipe_resource *resource, unsigned level, unsigned usage, const struct pipe_box *box) { struct r600_context *ctx = (struct r600_context *)ctx_; struct compute_memory_pool *pool = ctx->screen->global_pool; compute_memory_finalize_pending(pool, ctx_); assert(resource->target == PIPE_BUFFER); struct r600_context *rctx = (struct r600_context*)ctx_; struct pipe_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); transfer->resource = resource; transfer->level = level; transfer->usage = usage; transfer->box = *box; transfer->stride = 0; transfer->layer_stride = 0; transfer->data = NULL; /* Note strides are zero, this is ok for buffers, but not for * textures 2d & higher at least. */ return transfer; }
static void evergreen_set_global_binding( struct pipe_context *ctx_, unsigned first, unsigned n, struct pipe_resource **resources, uint32_t **handles) { struct r600_context *ctx = (struct r600_context *)ctx_; struct compute_memory_pool *pool = ctx->screen->global_pool; struct r600_resource_global **buffers = (struct r600_resource_global **)resources; COMPUTE_DBG("*** evergreen_set_global_binding first = %u n = %u\n", first, n); if (!resources) { /* XXX: Unset */ return; } compute_memory_finalize_pending(pool, ctx_); for (int i = 0; i < n; i++) { assert(resources[i]->target == PIPE_BUFFER); assert(resources[i]->bind & PIPE_BIND_GLOBAL); *(handles[i]) = buffers[i]->chunk->start_in_dw * 4; } evergreen_set_rat(ctx->cs_shader_state.shader, 0, pool->bo, 0, pool->size_in_dw * 4); evergreen_cs_set_vertex_buffer(ctx, 1, 0, (struct pipe_resource*)pool->bo); }
void *r600_compute_global_transfer_map( struct pipe_context *ctx_, struct pipe_resource *resource, unsigned level, unsigned usage, const struct pipe_box *box, struct pipe_transfer **ptransfer) { struct r600_context *rctx = (struct r600_context*)ctx_; struct compute_memory_pool *pool = rctx->screen->global_pool; struct r600_resource_global* buffer = (struct r600_resource_global*)resource; COMPUTE_DBG(rctx->screen, "* r600_compute_global_transfer_map()\n" "level = %u, usage = %u, box(x = %u, y = %u, z = %u " "width = %u, height = %u, depth = %u)\n", level, usage, box->x, box->y, box->z, box->width, box->height, box->depth); COMPUTE_DBG(rctx->screen, "Buffer id = %u offset = " "%u (box.x)\n", buffer->chunk->id, box->x); compute_memory_finalize_pending(pool, ctx_); assert(resource->target == PIPE_BUFFER); assert(resource->bind & PIPE_BIND_GLOBAL); assert(box->x >= 0); assert(box->y == 0); assert(box->z == 0); ///TODO: do it better, mapping is not possible if the pool is too big return pipe_buffer_map_range(ctx_, (struct pipe_resource*)buffer->chunk->pool->bo, box->x + (buffer->chunk->start_in_dw * 4), box->width, usage, ptransfer); }
static void evergreen_set_global_binding(struct pipe_context *ctx, unsigned first, unsigned n, struct pipe_resource **resources, uint32_t **handles) { struct r600_context *rctx = (struct r600_context *)ctx; struct compute_memory_pool *pool = rctx->screen->global_pool; struct r600_resource_global **buffers = (struct r600_resource_global **)resources; unsigned i; COMPUTE_DBG(rctx->screen, "*** evergreen_set_global_binding first = %u n = %u\n", first, n); if (!resources) { /* XXX: Unset */ return; } /* We mark these items for promotion to the pool if they * aren't already there */ for (i = first; i < first + n; i++) { struct compute_memory_item *item = buffers[i]->chunk; if (!is_item_in_pool(item)) buffers[i]->chunk->status |= ITEM_FOR_PROMOTING; } if (compute_memory_finalize_pending(pool, ctx) == -1) { /* XXX: Unset */ return; } for (i = first; i < first + n; i++) { uint32_t buffer_offset; uint32_t handle; assert(resources[i]->target == PIPE_BUFFER); assert(resources[i]->bind & PIPE_BIND_GLOBAL); buffer_offset = util_le32_to_cpu(*(handles[i])); handle = buffer_offset + buffers[i]->chunk->start_in_dw * 4; *(handles[i]) = util_cpu_to_le32(handle); } /* globals for writing */ evergreen_set_rat(rctx->cs_shader_state.shader, 0, pool->bo, 0, pool->size_in_dw * 4); /* globals for reading */ evergreen_cs_set_vertex_buffer(rctx, 1, 0, (struct pipe_resource*)pool->bo); /* constants for reading, LLVM puts them in text segment */ evergreen_cs_set_vertex_buffer(rctx, 2, 0, (struct pipe_resource*)rctx->cs_shader_state.shader->code_bo); }
void *r600_compute_global_transfer_map( struct pipe_context *ctx_, struct pipe_resource *resource, unsigned level, unsigned usage, const struct pipe_box *box, struct pipe_transfer **ptransfer) { struct r600_context *rctx = (struct r600_context*)ctx_; struct compute_memory_pool *pool = rctx->screen->global_pool; struct pipe_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); struct r600_resource_global* buffer = (struct r600_resource_global*)resource; uint32_t* map; compute_memory_finalize_pending(pool, ctx_); assert(resource->target == PIPE_BUFFER); COMPUTE_DBG(rctx->screen, "* r600_compute_global_get_transfer()\n" "level = %u, usage = %u, box(x = %u, y = %u, z = %u " "width = %u, height = %u, depth = %u)\n", level, usage, box->x, box->y, box->z, box->width, box->height, box->depth); transfer->resource = resource; transfer->level = level; transfer->usage = usage; transfer->box = *box; transfer->stride = 0; transfer->layer_stride = 0; assert(transfer->resource->target == PIPE_BUFFER); assert(transfer->resource->bind & PIPE_BIND_GLOBAL); assert(transfer->box.x >= 0); assert(transfer->box.y == 0); assert(transfer->box.z == 0); ///TODO: do it better, mapping is not possible if the pool is too big COMPUTE_DBG(rctx->screen, "* r600_compute_global_transfer_map()\n"); if (!(map = r600_buffer_mmap_sync_with_rings(rctx, buffer->chunk->pool->bo, transfer->usage))) { util_slab_free(&rctx->pool_transfers, transfer); return NULL; } *ptransfer = transfer; COMPUTE_DBG(rctx->screen, "Buffer: %p + %u (buffer offset in global memory) " "+ %u (box.x)\n", map, buffer->chunk->start_in_dw, transfer->box.x); return ((char*)(map + buffer->chunk->start_in_dw)) + transfer->box.x; }