/external/chromium_org/third_party/mesa/src/src/gallium/drivers/radeonsi/ |
H A D | r600_query.c | 28 struct r600_context *rctx = (struct r600_context *)ctx; local 30 return (struct pipe_query*)r600_context_query_create(rctx, query_type); 35 struct r600_context *rctx = (struct r600_context *)ctx; local 37 r600_context_query_destroy(rctx, (struct r600_query *)query); 42 struct r600_context *rctx = (struct r600_context *)ctx; local 47 r600_query_begin(rctx, (struct r600_query *)query); 48 LIST_ADDTAIL(&rquery->list, &rctx->active_query_list); 53 struct r600_context *rctx = (struct r600_context *)ctx; local 56 r600_query_end(rctx, rquery); 64 struct r600_context *rctx local 74 struct r600_context *rctx = (struct r600_context *)ctx; local 120 r600_init_query_functions(struct r600_context *rctx) argument [all...] |
H A D | r600_buffer.c | 55 struct r600_context *rctx = (struct r600_context*)ctx; local 56 struct pipe_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); 76 struct r600_context *rctx = (struct r600_context*)pipe; local 79 data = rctx->ws->buffer_map(rbuffer->cs_buf, rctx->cs, transfer->usage); 101 struct r600_context *rctx = (struct r600_context*)ctx; local 102 util_slab_free(&rctx->pool_transfers, transfer); 181 void r600_upload_index_buffer(struct r600_context *rctx, argument 184 u_upload_data(rctx->uploader, 0, count * ib->index_size, 188 void r600_upload_const_buffer(struct r600_context *rctx, struc argument [all...] |
H A D | radeonsi_pm4.c | 157 void si_pm4_free_state(struct r600_context *rctx, argument 164 if (idx != ~0 && rctx->emitted.array[idx] == state) { 165 rctx->emitted.array[idx] = NULL; 174 uint32_t si_pm4_sync_flags(struct r600_context *rctx) argument 179 struct si_pm4_state *state = rctx->queued.array[i]; 181 if (!state || rctx->emitted.array[i] == state) 189 unsigned si_pm4_dirty_dw(struct r600_context *rctx) argument 194 struct si_pm4_state *state = rctx->queued.array[i]; 196 if (!state || rctx->emitted.array[i] == state) 205 void si_pm4_emit(struct r600_context *rctx, struc argument 222 si_pm4_emit_dirty(struct r600_context *rctx) argument 235 si_pm4_reset_emitted(struct r600_context *rctx) argument [all...] |
H A D | r600_texture.c | 402 struct r600_context *rctx = (struct r600_context *)ctx; local 424 if (!(map = rctx->ws->buffer_map(buf, rctx->cs, transfer->usage))) { 435 struct r600_context *rctx = (struct r600_context*)ctx; local 449 rctx->ws->buffer_unmap(buf);
|
H A D | radeonsi_pipe.c | 56 static struct r600_fence *r600_create_fence(struct r600_context *rctx) argument 58 struct r600_screen *rscreen = rctx->screen; 72 rscreen->fences.data = rctx->ws->buffer_map(rscreen->fences.bo->cs_buf, 73 rctx->cs, 120 si_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1); 123 fence->sleep_bo = si_resource_create_custom(&rctx->screen->screen, PIPE_USAGE_STAGING, 1); 126 r600_context_bo_reloc(rctx, fence->sleep_bo, RADEON_USAGE_READWRITE); 137 struct r600_context *rctx = (struct r600_context *)ctx; local 143 *rfence = r600_create_fence(rctx); 146 if (rctx 173 struct r600_context *rctx = (struct r600_context *)context; local 192 struct r600_context *rctx = CALLOC_STRUCT(r600_context); local [all...] |
H A D | si_state_draw.c | 42 struct r600_context *rctx = (struct r600_context *)ctx; local 51 si_pm4_delete_state(rctx, vs, shader->pm4); 95 si_pm4_bind_state(rctx, vs, shader->pm4); 100 struct r600_context *rctx = (struct r600_context *)ctx; local 112 si_pm4_delete_state(rctx, ps, shader->pm4); 123 rctx->queued.named.rasterizer->flatshade)) 211 shader->sprite_coord_enable = rctx->sprite_coord_enable; 212 si_pm4_bind_state(rctx, ps, shader->pm4); 244 static bool si_update_draw_info_state(struct r600_context *rctx, argument 277 rctx 298 si_update_alpha_ref(struct r600_context *rctx) argument 315 si_update_spi_map(struct r600_context *rctx) argument 359 si_update_derived_state(struct r600_context *rctx) argument 403 si_vertex_buffer_update(struct r600_context *rctx) argument 456 si_state_draw(struct r600_context *rctx, const struct pipe_draw_info *info, const struct pipe_index_buffer *ib) argument 516 struct r600_context *rctx = (struct r600_context *)ctx; local [all...] |
H A D | r600_blit.c | 47 struct r600_context *rctx = (struct r600_context *)ctx; local 49 r600_context_queries_suspend(rctx); 51 util_blitter_save_blend(rctx->blitter, rctx->queued.named.blend); 52 util_blitter_save_depth_stencil_alpha(rctx->blitter, rctx->queued.named.dsa); 53 util_blitter_save_stencil_ref(rctx->blitter, &rctx->stencil_ref); 54 util_blitter_save_rasterizer(rctx->blitter, rctx 90 struct r600_context *rctx = (struct r600_context *)ctx; local 118 struct r600_context *rctx = (struct r600_context *)ctx; local 156 si_flush_depth_textures(struct r600_context *rctx) argument 198 struct r600_context *rctx = (struct r600_context *)ctx; local 214 struct r600_context *rctx = (struct r600_context *)ctx; local 230 struct r600_context *rctx = (struct r600_context *)ctx; local 249 struct r600_context *rctx = (struct r600_context *)ctx; local 377 si_init_blit_functions(struct r600_context *rctx) argument [all...] |
/external/mesa3d/src/gallium/drivers/radeonsi/ |
H A D | r600_query.c | 28 struct r600_context *rctx = (struct r600_context *)ctx; local 30 return (struct pipe_query*)r600_context_query_create(rctx, query_type); 35 struct r600_context *rctx = (struct r600_context *)ctx; local 37 r600_context_query_destroy(rctx, (struct r600_query *)query); 42 struct r600_context *rctx = (struct r600_context *)ctx; local 47 r600_query_begin(rctx, (struct r600_query *)query); 48 LIST_ADDTAIL(&rquery->list, &rctx->active_query_list); 53 struct r600_context *rctx = (struct r600_context *)ctx; local 56 r600_query_end(rctx, rquery); 64 struct r600_context *rctx local 74 struct r600_context *rctx = (struct r600_context *)ctx; local 120 r600_init_query_functions(struct r600_context *rctx) argument [all...] |
H A D | r600_buffer.c | 55 struct r600_context *rctx = (struct r600_context*)ctx; local 56 struct pipe_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); 76 struct r600_context *rctx = (struct r600_context*)pipe; local 79 data = rctx->ws->buffer_map(rbuffer->cs_buf, rctx->cs, transfer->usage); 101 struct r600_context *rctx = (struct r600_context*)ctx; local 102 util_slab_free(&rctx->pool_transfers, transfer); 181 void r600_upload_index_buffer(struct r600_context *rctx, argument 184 u_upload_data(rctx->uploader, 0, count * ib->index_size, 188 void r600_upload_const_buffer(struct r600_context *rctx, struc argument [all...] |
H A D | radeonsi_pm4.c | 157 void si_pm4_free_state(struct r600_context *rctx, argument 164 if (idx != ~0 && rctx->emitted.array[idx] == state) { 165 rctx->emitted.array[idx] = NULL; 174 uint32_t si_pm4_sync_flags(struct r600_context *rctx) argument 179 struct si_pm4_state *state = rctx->queued.array[i]; 181 if (!state || rctx->emitted.array[i] == state) 189 unsigned si_pm4_dirty_dw(struct r600_context *rctx) argument 194 struct si_pm4_state *state = rctx->queued.array[i]; 196 if (!state || rctx->emitted.array[i] == state) 205 void si_pm4_emit(struct r600_context *rctx, struc argument 222 si_pm4_emit_dirty(struct r600_context *rctx) argument 235 si_pm4_reset_emitted(struct r600_context *rctx) argument [all...] |
H A D | r600_texture.c | 402 struct r600_context *rctx = (struct r600_context *)ctx; local 424 if (!(map = rctx->ws->buffer_map(buf, rctx->cs, transfer->usage))) { 435 struct r600_context *rctx = (struct r600_context*)ctx; local 449 rctx->ws->buffer_unmap(buf);
|
H A D | radeonsi_pipe.c | 56 static struct r600_fence *r600_create_fence(struct r600_context *rctx) argument 58 struct r600_screen *rscreen = rctx->screen; 72 rscreen->fences.data = rctx->ws->buffer_map(rscreen->fences.bo->cs_buf, 73 rctx->cs, 120 si_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1); 123 fence->sleep_bo = si_resource_create_custom(&rctx->screen->screen, PIPE_USAGE_STAGING, 1); 126 r600_context_bo_reloc(rctx, fence->sleep_bo, RADEON_USAGE_READWRITE); 137 struct r600_context *rctx = (struct r600_context *)ctx; local 143 *rfence = r600_create_fence(rctx); 146 if (rctx 173 struct r600_context *rctx = (struct r600_context *)context; local 192 struct r600_context *rctx = CALLOC_STRUCT(r600_context); local [all...] |
H A D | si_state_draw.c | 42 struct r600_context *rctx = (struct r600_context *)ctx; local 51 si_pm4_delete_state(rctx, vs, shader->pm4); 95 si_pm4_bind_state(rctx, vs, shader->pm4); 100 struct r600_context *rctx = (struct r600_context *)ctx; local 112 si_pm4_delete_state(rctx, ps, shader->pm4); 123 rctx->queued.named.rasterizer->flatshade)) 211 shader->sprite_coord_enable = rctx->sprite_coord_enable; 212 si_pm4_bind_state(rctx, ps, shader->pm4); 244 static bool si_update_draw_info_state(struct r600_context *rctx, argument 277 rctx 298 si_update_alpha_ref(struct r600_context *rctx) argument 315 si_update_spi_map(struct r600_context *rctx) argument 359 si_update_derived_state(struct r600_context *rctx) argument 403 si_vertex_buffer_update(struct r600_context *rctx) argument 456 si_state_draw(struct r600_context *rctx, const struct pipe_draw_info *info, const struct pipe_index_buffer *ib) argument 516 struct r600_context *rctx = (struct r600_context *)ctx; local [all...] |
H A D | r600_blit.c | 47 struct r600_context *rctx = (struct r600_context *)ctx; local 49 r600_context_queries_suspend(rctx); 51 util_blitter_save_blend(rctx->blitter, rctx->queued.named.blend); 52 util_blitter_save_depth_stencil_alpha(rctx->blitter, rctx->queued.named.dsa); 53 util_blitter_save_stencil_ref(rctx->blitter, &rctx->stencil_ref); 54 util_blitter_save_rasterizer(rctx->blitter, rctx 90 struct r600_context *rctx = (struct r600_context *)ctx; local 118 struct r600_context *rctx = (struct r600_context *)ctx; local 156 si_flush_depth_textures(struct r600_context *rctx) argument 198 struct r600_context *rctx = (struct r600_context *)ctx; local 214 struct r600_context *rctx = (struct r600_context *)ctx; local 230 struct r600_context *rctx = (struct r600_context *)ctx; local 249 struct r600_context *rctx = (struct r600_context *)ctx; local 377 si_init_blit_functions(struct r600_context *rctx) argument [all...] |
/external/chromium_org/third_party/mesa/src/src/gallium/drivers/r600/ |
H A D | r600_buffer.c | 46 struct r600_context *rctx = (struct r600_context*)ctx; local 47 struct r600_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); 67 static void r600_set_constants_dirty_if_bound(struct r600_context *rctx, argument 82 r600_constant_buffers_dirty(rctx, state); 90 struct r600_context *rctx = (struct r600_context*)pipe; local 98 if (rctx->ws->cs_is_buffer_referenced(rctx->cs, rbuffer->cs_buf, RADEON_USAGE_READWRITE) || 99 rctx->ws->buffer_is_busy(rbuffer->buf, RADEON_USAGE_READWRITE)) { 107 r600_init_resource(rctx->screen, rbuffer, rbuffer->b.b.width0, 4096, 112 mask = rctx 181 struct r600_context *rctx = (struct r600_context*)ctx; local [all...] |
H A D | r600_pipe.c | 39 static struct r600_fence *r600_create_fence(struct r600_context *rctx) argument 41 struct r600_screen *rscreen = rctx->screen; 55 rscreen->fences.data = rctx->ws->buffer_map(rscreen->fences.bo->cs_buf, 56 rctx->cs, 103 r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1); 107 pipe_buffer_create(&rctx->screen->screen, PIPE_BIND_CUSTOM, 110 r600_context_bo_reloc(rctx, fence->sleep_bo, RADEON_USAGE_READWRITE); 121 struct r600_context *rctx = (struct r600_context *)ctx; local 127 *rfence = r600_create_fence(rctx); 130 if (rctx 157 struct r600_context *rctx = (struct r600_context *)context; local 205 struct r600_context *rctx = CALLOC_STRUCT(r600_context); local [all...] |
H A D | r600_query.c | 253 struct r600_context *rctx = (struct r600_context *)ctx; local 266 query->result_size = 16 * rctx->max_db; 291 query->buffer.buf = r600_new_query_buffer(rctx, query_type); 316 static void r600_update_occlusion_query_state(struct r600_context *rctx, argument 323 rctx->num_occlusion_queries += diff; 324 assert(rctx->num_occlusion_queries >= 0); 326 enable = rctx->num_occlusion_queries != 0; 328 if (rctx->db_misc_state.occlusion_query_enabled != enable) { 329 rctx->db_misc_state.occlusion_query_enabled = enable; 330 r600_atom_dirty(rctx, 337 struct r600_context *rctx = (struct r600_context *)ctx; local 377 struct r600_context *rctx = (struct r600_context *)ctx; local 502 struct r600_context *rctx = (struct r600_context *)ctx; local 526 struct r600_context *rctx = (struct r600_context *)ctx; local 607 r600_init_query_functions(struct r600_context *rctx) argument [all...] |
/external/mesa3d/src/gallium/drivers/r600/ |
H A D | r600_buffer.c | 46 struct r600_context *rctx = (struct r600_context*)ctx; local 47 struct r600_transfer *transfer = util_slab_alloc(&rctx->pool_transfers); 67 static void r600_set_constants_dirty_if_bound(struct r600_context *rctx, argument 82 r600_constant_buffers_dirty(rctx, state); 90 struct r600_context *rctx = (struct r600_context*)pipe; local 98 if (rctx->ws->cs_is_buffer_referenced(rctx->cs, rbuffer->cs_buf, RADEON_USAGE_READWRITE) || 99 rctx->ws->buffer_is_busy(rbuffer->buf, RADEON_USAGE_READWRITE)) { 107 r600_init_resource(rctx->screen, rbuffer, rbuffer->b.b.width0, 4096, 112 mask = rctx 181 struct r600_context *rctx = (struct r600_context*)ctx; local [all...] |
H A D | r600_pipe.c | 39 static struct r600_fence *r600_create_fence(struct r600_context *rctx) argument 41 struct r600_screen *rscreen = rctx->screen; 55 rscreen->fences.data = rctx->ws->buffer_map(rscreen->fences.bo->cs_buf, 56 rctx->cs, 103 r600_context_emit_fence(rctx, rscreen->fences.bo, fence->index, 1); 107 pipe_buffer_create(&rctx->screen->screen, PIPE_BIND_CUSTOM, 110 r600_context_bo_reloc(rctx, fence->sleep_bo, RADEON_USAGE_READWRITE); 121 struct r600_context *rctx = (struct r600_context *)ctx; local 127 *rfence = r600_create_fence(rctx); 130 if (rctx 157 struct r600_context *rctx = (struct r600_context *)context; local 205 struct r600_context *rctx = CALLOC_STRUCT(r600_context); local [all...] |
H A D | r600_query.c | 253 struct r600_context *rctx = (struct r600_context *)ctx; local 266 query->result_size = 16 * rctx->max_db; 291 query->buffer.buf = r600_new_query_buffer(rctx, query_type); 316 static void r600_update_occlusion_query_state(struct r600_context *rctx, argument 323 rctx->num_occlusion_queries += diff; 324 assert(rctx->num_occlusion_queries >= 0); 326 enable = rctx->num_occlusion_queries != 0; 328 if (rctx->db_misc_state.occlusion_query_enabled != enable) { 329 rctx->db_misc_state.occlusion_query_enabled = enable; 330 r600_atom_dirty(rctx, 337 struct r600_context *rctx = (struct r600_context *)ctx; local 377 struct r600_context *rctx = (struct r600_context *)ctx; local 502 struct r600_context *rctx = (struct r600_context *)ctx; local 526 struct r600_context *rctx = (struct r600_context *)ctx; local 607 r600_init_query_functions(struct r600_context *rctx) argument [all...] |
/external/chromium_org/third_party/boringssl/src/crypto/evp/ |
H A D | evp_ctx.c | 160 EVP_PKEY_CTX *rctx; local 166 rctx = OPENSSL_malloc(sizeof(EVP_PKEY_CTX)); 167 if (!rctx) { 171 memset(rctx, 0, sizeof(EVP_PKEY_CTX)); 173 rctx->pmeth = pctx->pmeth; 174 rctx->engine = pctx->engine; 175 rctx->operation = pctx->operation; 178 rctx->pkey = EVP_PKEY_dup(pctx->pkey); 182 rctx->peerkey = EVP_PKEY_dup(pctx->peerkey); 185 if (pctx->pmeth->copy(rctx, pct [all...] |
H A D | p_rsa.c | 91 RSA_PKEY_CTX *rctx; local 92 rctx = OPENSSL_malloc(sizeof(RSA_PKEY_CTX)); 93 if (!rctx) { 96 memset(rctx, 0, sizeof(RSA_PKEY_CTX)); 98 rctx->nbits = 2048; 99 rctx->pad_mode = RSA_PKCS1_PADDING; 100 rctx->saltlen = -2; 102 ctx->data = rctx; 140 RSA_PKEY_CTX *rctx = ctx->data; local 142 if (rctx 171 RSA_PKEY_CTX *rctx = ctx->data; local 227 RSA_PKEY_CTX *rctx = ctx->data; local 265 RSA_PKEY_CTX *rctx = ctx->data; local 297 RSA_PKEY_CTX *rctx = ctx->data; local 360 RSA_PKEY_CTX *rctx = ctx->data; local 494 RSA_PKEY_CTX *rctx = ctx->data; local [all...] |
/external/openssl/crypto/ocsp/ |
H A D | ocsp_ht.c | 112 void OCSP_REQ_CTX_free(OCSP_REQ_CTX *rctx) argument 114 if (rctx->mem) 115 BIO_free(rctx->mem); 116 if (rctx->iobuf) 117 OPENSSL_free(rctx->iobuf); 118 OPENSSL_free(rctx); 121 int OCSP_REQ_CTX_set1_req(OCSP_REQ_CTX *rctx, OCSP_REQUEST *req) argument 126 if (BIO_printf(rctx->mem, req_hdr, i2d_OCSP_REQUEST(req, NULL)) <= 0) 128 if (i2d_OCSP_REQUEST_bio(rctx->mem, req) <= 0) 130 rctx 135 OCSP_REQ_CTX_add1_header(OCSP_REQ_CTX *rctx, const char *name, const char *value) argument 159 OCSP_REQ_CTX *rctx; local 268 OCSP_sendreq_nbio(OCSP_RESPONSE **presp, OCSP_REQ_CTX *rctx) argument [all...] |
/external/openssl/crypto/rsa/ |
H A D | rsa_pmeth.c | 98 RSA_PKEY_CTX *rctx; local 99 rctx = OPENSSL_malloc(sizeof(RSA_PKEY_CTX)); 100 if (!rctx) 102 rctx->nbits = 1024; 103 rctx->pub_exp = NULL; 104 rctx->pad_mode = RSA_PKCS1_PADDING; 105 rctx->md = NULL; 106 rctx->mgf1md = NULL; 107 rctx->tbuf = NULL; 109 rctx 149 RSA_PKEY_CTX *rctx = ctx->data; local 168 RSA_PKEY_CTX *rctx = ctx->data; local 189 RSA_PKEY_CTX *rctx = ctx->data; local 287 RSA_PKEY_CTX *rctx = ctx->data; local 343 RSA_PKEY_CTX *rctx = ctx->data; local 421 RSA_PKEY_CTX *rctx = ctx->data; local 435 RSA_PKEY_CTX *rctx = ctx->data; local 472 RSA_PKEY_CTX *rctx = ctx->data; local 661 RSA_PKEY_CTX *rctx = ctx->data; local [all...] |
/external/openssl/crypto/evp/ |
H A D | pmeth_lib.c | 309 EVP_PKEY_CTX *rctx; local 320 rctx = OPENSSL_malloc(sizeof(EVP_PKEY_CTX)); 321 if (!rctx) 324 rctx->pmeth = pctx->pmeth; 326 rctx->engine = pctx->engine; 332 rctx->pkey = pctx->pkey; 337 rctx->peerkey = pctx->peerkey; 339 rctx->data = NULL; 340 rctx->app_data = NULL; 341 rctx [all...] |