/external/mesa3d/src/gallium/drivers/freedreno/ |
H A D | freedreno_batch.c | 38 batch_init(struct fd_batch *batch) argument 40 struct fd_context *ctx = batch->ctx; 44 util_queue_fence_init(&batch->flush_fence); 56 batch->draw = fd_ringbuffer_new(ctx->screen->pipe, size); 57 batch->binning = fd_ringbuffer_new(ctx->screen->pipe, size); 58 batch->gmem = fd_ringbuffer_new(ctx->screen->pipe, size); 60 fd_ringbuffer_set_parent(batch->gmem, NULL); 61 fd_ringbuffer_set_parent(batch->draw, batch->gmem); 62 fd_ringbuffer_set_parent(batch 92 struct fd_batch *batch = CALLOC_STRUCT(fd_batch); local 111 batch_fini(struct fd_batch *batch) argument 139 batch_flush_reset_dependencies(struct fd_batch *batch, bool flush) argument 154 batch_reset_resources_locked(struct fd_batch *batch) argument 171 batch_reset_resources(struct fd_batch *batch) argument 179 batch_reset(struct fd_batch *batch) argument 193 fd_batch_reset(struct fd_batch *batch) argument 200 __fd_batch_destroy(struct fd_batch *batch) argument 223 __fd_batch_describe(char* buf, const struct fd_batch *batch) argument 229 fd_batch_sync(struct fd_batch *batch) argument 239 struct fd_batch *batch = job; local 248 struct fd_batch *batch = job; local 253 batch_flush(struct fd_batch *batch) argument 298 fd_batch_flush(struct fd_batch *batch, bool sync) argument 314 batch_depends_on(struct fd_batch *batch, struct fd_batch *other) argument 330 batch_add_dep(struct fd_batch *batch, struct fd_batch *dep) argument 352 fd_batch_resource_used(struct fd_batch *batch, struct fd_resource *rsc, bool write) argument 400 fd_batch_check_size(struct fd_batch *batch) argument 415 fd_wfi(struct fd_batch *batch, struct fd_ringbuffer *ring) argument [all...] |
H A D | freedreno_batch.h | 62 /* A batch tracks everything about a cmdstream batch/submit, including the 119 unsigned num_draws; /* number of draws in current batch */ 122 * batch. Used at the tile rendering step (fd_gmem_render_tiles(), 161 /* next sample offset.. incremented for each sample in the batch/ 171 /* which sample providers were active in the current batch: */ 179 /* list of samples in current batch: */ 188 /* Set of resources used by currently-unsubmitted batch (read or 193 /** key in batch-cache (if not null): */ 203 void fd_batch_reset(struct fd_batch *batch); 225 fd_batch_reference(struct fd_batch **ptr, struct fd_batch *batch) argument 240 fd_batch_reference_locked(struct fd_batch **ptr, struct fd_batch *batch) argument 262 fd_reset_wfi(struct fd_batch *batch) argument 272 fd_event_write(struct fd_batch *batch, struct fd_ringbuffer *ring, enum vgt_event_type evt) argument [all...] |
H A D | freedreno_gmem.c | 107 calculate_tiles(struct fd_batch *batch) argument 109 struct fd_context *ctx = batch->ctx; 111 struct pipe_scissor_state *scissor = &batch->max_scissor; 112 struct pipe_framebuffer_state *pfb = &batch->framebuffer; 123 bool has_zs = !!(batch->resolve & (FD_BUFFER_DEPTH | FD_BUFFER_STENCIL)); 310 render_tiles(struct fd_batch *batch) argument 312 struct fd_context *ctx = batch->ctx; 316 ctx->emit_tile_init(batch); 318 if (batch->restore) 327 ctx->emit_tile_prep(batch, til 350 render_sysmem(struct fd_batch *batch) argument 367 flush_ring(struct fd_batch *batch) argument 381 fd_gmem_render_tiles(struct fd_batch *batch) argument 427 fd_gmem_render_noop(struct fd_batch *batch) argument 458 fd_gmem_needs_restore(struct fd_batch *batch, struct fd_tile *tile, uint32_t buffers) argument [all...] |
H A D | freedreno_query_hw.c | 64 get_sample(struct fd_batch *batch, struct fd_ringbuffer *ring, argument 67 struct fd_context *ctx = batch->ctx; 73 if (!batch->sample_cache[idx]) { 75 ctx->sample_providers[idx]->get_sample(batch, ring); 76 fd_hw_sample_reference(ctx, &batch->sample_cache[idx], new_samp); 77 util_dynarray_append(&batch->samples, struct fd_hw_sample *, new_samp); 78 batch->needs_flush = true; 81 fd_hw_sample_reference(ctx, &samp, batch->sample_cache[idx]); 87 clear_sample_cache(struct fd_batch *batch) argument 91 for (i = 0; i < ARRAY_SIZE(batch 103 resume_query(struct fd_batch *batch, struct fd_hw_query *hq, struct fd_ringbuffer *ring) argument 119 pause_query(struct fd_batch *batch, struct fd_hw_query *hq, struct fd_ringbuffer *ring) argument 160 struct fd_batch *batch = ctx->batch; local 186 struct fd_batch *batch = ctx->batch; local 340 fd_hw_sample_init(struct fd_batch *batch, uint32_t size) argument 386 fd_hw_query_prepare(struct fd_batch *batch, uint32_t num_tiles) argument 408 fd_hw_query_prepare_tile(struct fd_batch *batch, uint32_t n, struct fd_ringbuffer *ring) argument 424 fd_hw_query_set_stage(struct fd_batch *batch, struct fd_ringbuffer *ring, enum fd_render_stage stage) argument 458 fd_hw_query_enable(struct fd_batch *batch, struct fd_ringbuffer *ring) argument [all...] |
H A D | freedreno_draw.c | 43 resource_read(struct fd_batch *batch, struct pipe_resource *prsc) argument 47 fd_batch_resource_used(batch, fd_resource(prsc), false); 51 resource_written(struct fd_batch *batch, struct pipe_resource *prsc) argument 55 fd_batch_resource_used(batch, fd_resource(prsc), true); 62 struct fd_batch *batch = ctx->batch; local 63 struct pipe_framebuffer_state *pfb = &batch->framebuffer; 88 fd_batch_reset(batch); 92 batch->blit = ctx->in_blit; 93 batch 284 struct fd_batch *batch = ctx->batch; local [all...] |
H A D | freedreno_batch_cache.c | 39 * The batch cache provides lookup for mapping pipe_framebuffer_state 40 * to a batch. 54 * Each batch needs to hold a reference to each resource it depends on (ie. 59 * When a resource is destroyed, we need to remove entries in the batch 64 * When a batch has weak reference to no more resources (ie. all the 65 * surfaces it rendered to are destroyed) the batch can be destroyed. 68 * surfaces are destroyed before the batch is submitted. 70 * If (for example), batch writes to zsbuf but that surface is destroyed 71 * before batch is submitted, we can skip gmem2mem (but still need to 72 * alloc gmem space as before. If the batch depende 136 struct fd_batch *batch = NULL; local 159 struct fd_batch *batch; local 172 fd_bc_invalidate_batch(struct fd_batch *batch, bool destroy) argument 208 struct fd_batch *batch; local 233 struct fd_batch *batch; local 315 struct fd_batch *batch = NULL; local [all...] |
H A D | freedreno_query_hw.h | 80 * at least once before the first ->get_sample() in a batch. 87 struct fd_hw_sample *(*get_sample)(struct fd_batch *batch, 106 * the sample is constructed, but when the batch is submitted. 130 struct list_head list; /* list-node in batch->active_queries */ 143 struct fd_hw_sample * fd_hw_sample_init(struct fd_batch *batch, uint32_t size); 146 void fd_hw_query_prepare(struct fd_batch *batch, uint32_t num_tiles); 147 void fd_hw_query_prepare_tile(struct fd_batch *batch, uint32_t n, 149 void fd_hw_query_set_stage(struct fd_batch *batch, 151 void fd_hw_query_enable(struct fd_batch *batch, struct fd_ringbuffer *ring);
|
/external/mesa3d/src/gallium/drivers/i915/ |
H A D | i915_batchbuffer.h | 37 i915_winsys_batchbuffer_space(struct i915_winsys_batchbuffer *batch) argument 39 return batch->size - (batch->ptr - batch->map); 43 i915_winsys_batchbuffer_check(struct i915_winsys_batchbuffer *batch, argument 46 return dwords * 4 <= i915_winsys_batchbuffer_space(batch); 50 i915_winsys_batchbuffer_dword_unchecked(struct i915_winsys_batchbuffer *batch, argument 53 *(unsigned *)batch->ptr = dword; 54 batch->ptr += 4; 58 i915_winsys_batchbuffer_float(struct i915_winsys_batchbuffer *batch, argument 68 i915_winsys_batchbuffer_dword(struct i915_winsys_batchbuffer *batch, unsigned dword) argument 76 i915_winsys_batchbuffer_write(struct i915_winsys_batchbuffer *batch, void *data, size_t size) argument 87 i915_winsys_validate_buffers(struct i915_winsys_batchbuffer *batch, struct i915_winsys_buffer **buffers, int num_of_buffers) argument 95 i915_winsys_batchbuffer_reloc(struct i915_winsys_batchbuffer *batch, struct i915_winsys_buffer *buffer, enum i915_winsys_buffer_usage usage, size_t offset, boolean fenced) argument [all...] |
H A D | i915_batch.h | 36 (i915_winsys_batchbuffer_check(i915->batch, dwords)) 39 i915_winsys_batchbuffer_dword(i915->batch, dword) 42 i915_winsys_batchbuffer_float(i915->batch, f) 45 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, false) 48 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, true)
|
H A D | i915_flush.c | 48 if (!i915->batch) 54 if (!fence && (i915->batch->map == i915->batch->ptr)) { 79 struct i915_winsys_batchbuffer *batch = i915->batch; local 81 batch->iws->batchbuffer_flush(batch, fence, flags);
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
H A D | batch_scheduler_test.cc | 41 Batch<FakeTask> batch; local 43 EXPECT_EQ(0, batch.num_tasks()); 44 EXPECT_TRUE(batch.empty()); 45 EXPECT_EQ(0, batch.size()); 46 EXPECT_FALSE(batch.IsClosed()); 49 batch.AddTask(std::unique_ptr<FakeTask>(task0)); 51 EXPECT_EQ(1, batch.num_tasks()); 52 EXPECT_FALSE(batch.empty()); 53 EXPECT_EQ(task0->size(), batch.size()); 54 EXPECT_EQ(task0->size(), batch 84 Batch<FakeTask> batch; local 98 Batch<FakeTask>* batch = new Batch<FakeTask>; local [all...] |
/external/mesa3d/src/mesa/drivers/dri/i965/ |
H A D | intel_batchbuffer.h | 14 * Number of bytes to reserve for commands necessary to complete a batch. 18 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes) 42 void intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, 44 void intel_batchbuffer_free(struct intel_batchbuffer *batch); 66 uint32_t intel_batchbuffer_reloc(struct intel_batchbuffer *batch, 72 uint64_t intel_batchbuffer_reloc64(struct intel_batchbuffer *batch, 79 #define USED_BATCH(batch) ((uintptr_t)((batch).map_next - (batch).map)) 98 intel_batchbuffer_space(struct intel_batchbuffer *batch) argument 106 intel_batchbuffer_emit_dword(struct intel_batchbuffer *batch, GLuint dword) argument 116 intel_batchbuffer_emit_float(struct intel_batchbuffer *batch, float f) argument 136 struct intel_batchbuffer *batch = &brw->batch; local [all...] |
H A D | genX_blorp_exec.c | 38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n) argument 40 assert(batch->blorp->driver_ctx == batch->driver_batch); 41 struct brw_context *brw = batch->driver_batch; 44 uint32_t *map = brw->batch.map_next; 45 brw->batch.map_next += n; 51 blorp_emit_reloc(struct blorp_batch *batch, argument 54 assert(batch->blorp->driver_ctx == batch->driver_batch); 55 struct brw_context *brw = batch 72 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, struct blorp_address address, uint32_t delta) argument 93 blorp_alloc_dynamic_state(struct blorp_batch *batch, enum aub_state_struct_type type, uint32_t size, uint32_t alignment, uint32_t *offset) argument 106 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries, unsigned state_size, unsigned state_alignment, uint32_t *bt_offset, uint32_t *surface_offsets, void **surface_maps) argument 127 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size, struct blorp_address *addr) argument 160 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size) argument 168 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size) argument 187 blorp_exec(struct blorp_batch *batch, const struct blorp_params *params) argument [all...] |
H A D | intel_batchbuffer.c | 39 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, 43 intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, argument 46 intel_batchbuffer_reset(batch, bufmgr, has_llc); 49 batch->cpu_map = malloc(BATCH_SZ); 50 batch->map = batch->cpu_map; 51 batch->map_next = batch->cpu_map; 56 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, argument 59 if (batch 109 intel_batchbuffer_free(struct intel_batchbuffer *batch) argument 146 struct intel_batchbuffer *batch = &brw->batch; local 324 struct intel_batchbuffer *batch = &brw->batch; local 442 intel_batchbuffer_reloc(struct intel_batchbuffer *batch, drm_intel_bo *buffer, uint32_t offset, uint32_t read_domains, uint32_t write_domain, uint32_t delta) argument 463 intel_batchbuffer_reloc64(struct intel_batchbuffer *batch, drm_intel_bo *buffer, uint32_t offset, uint32_t read_domains, uint32_t write_domain, uint32_t delta) argument [all...] |
H A D | brw_state_batch.c | 44 struct intel_batchbuffer *batch = &brw->batch; local 52 batch->bo->size / 32); 76 * Generate a set of aub file annotations for the current batch buffer, and 79 * The "used" section of the batch buffer (the portion containing batch 81 * batch buffer (which contains data structures pointed to by batch commands) 91 4 * USED_BATCH(brw->batch)); 102 drm_intel_bufmgr_gem_set_aub_annotations(brw->batch 129 struct intel_batchbuffer *batch = &brw->batch; local [all...] |
H A D | brw_blorp.h | 74 void gen6_blorp_exec(struct blorp_batch *batch, 76 void gen7_blorp_exec(struct blorp_batch *batch, 78 void gen75_blorp_exec(struct blorp_batch *batch, 80 void gen8_blorp_exec(struct blorp_batch *batch, 82 void gen9_blorp_exec(struct blorp_batch *batch,
|
/external/mesa3d/src/gallium/winsys/i915/drm/ |
H A D | i915_drm_batchbuffer.c | 30 i915_drm_batchbuffer(struct i915_winsys_batchbuffer *batch) argument 32 return (struct i915_drm_batchbuffer *)batch; 36 i915_drm_batchbuffer_reset(struct i915_drm_batchbuffer *batch) argument 38 struct i915_drm_winsys *idws = i915_drm_winsys(batch->base.iws); 40 if (batch->bo) 41 drm_intel_bo_unreference(batch->bo); 42 batch->bo = drm_intel_bo_alloc(idws->gem_manager, 44 batch->actual_size, 47 memset(batch->base.map, 0, batch 57 struct i915_drm_batchbuffer *batch = CALLOC_STRUCT(i915_drm_batchbuffer); local 75 i915_drm_batchbuffer_validate_buffers(struct i915_winsys_batchbuffer *batch, struct i915_winsys_buffer **buffer, int num_of_buffers) argument 100 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local 165 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local 221 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local [all...] |
/external/mesa3d/src/mesa/drivers/dri/i915/ |
H A D | intel_batchbuffer.c | 43 intel->batch.cpu_map = malloc(intel->maxBatchSize); 44 intel->batch.map = intel->batch.cpu_map; 50 if (intel->batch.last_bo != NULL) { 51 drm_intel_bo_unreference(intel->batch.last_bo); 52 intel->batch.last_bo = NULL; 54 intel->batch.last_bo = intel->batch.bo; 56 intel->batch.bo = drm_intel_bo_alloc(intel->bufmgr, "batchbuffer", 59 intel->batch 75 struct intel_batchbuffer *batch = &intel->batch; local 116 struct intel_batchbuffer *batch = &intel->batch; local [all...] |
H A D | intel_batchbuffer.h | 15 * Number of bytes to reserve for commands necessary to complete a batch. 19 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes) 76 return (intel->batch.bo->size - intel->batch.reserved_space) 77 - intel->batch.used*4; 87 intel->batch.map[intel->batch.used++] = dword; 112 intel->batch.emit = intel->batch.used; 114 intel->batch 122 struct intel_batchbuffer *batch = &intel->batch; local [all...] |
/external/mesa3d/src/intel/vulkan/ |
H A D | genX_blorp_exec.c | 38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n) argument 40 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch; 41 return anv_batch_emit_dwords(&cmd_buffer->batch, n); 45 blorp_emit_reloc(struct blorp_batch *batch, argument 48 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch; 49 assert(cmd_buffer->batch.start <= location && 50 location < cmd_buffer->batch.end); 51 return anv_batch_emit_reloc(&cmd_buffer->batch, location, 56 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, argument 59 struct anv_cmd_buffer *cmd_buffer = batch 65 blorp_alloc_dynamic_state(struct blorp_batch *batch, enum aub_state_struct_type type, uint32_t size, uint32_t alignment, uint32_t *offset) argument 81 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries, unsigned state_size, unsigned state_alignment, uint32_t *bt_offset, uint32_t *surface_offsets, void **surface_maps) argument 109 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size, struct blorp_address *addr) argument 138 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size) argument 146 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size) argument 164 blorp_exec(struct blorp_batch *batch, const struct blorp_params *params) argument [all...] |
/external/jsoncpp/src/lib_json/ |
H A D | json_batchallocator.h | 18 * This memory allocator allocates memory for a batch of object (specified by 44 for (BatchInfo* batch = batches_; batch;) { 45 BatchInfo* nextBatch = batch->next_; 46 free(batch); 47 batch = nextBatch; 66 if (!currentBatch_) // no free batch found, allocate a new one 103 BatchInfo* batch = static_cast<BatchInfo*>(malloc(mallocSize)); local 104 batch->next_ = 0; 105 batch [all...] |
/external/mesa3d/src/intel/blorp/ |
H A D | blorp_genX_exec.h | 48 blorp_emit_dwords(struct blorp_batch *batch, unsigned n); 51 blorp_emit_reloc(struct blorp_batch *batch, 55 blorp_alloc_dynamic_state(struct blorp_batch *batch, 61 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size, 65 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries, 71 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size); 74 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, 78 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size); 85 _blorp_combine_address(struct blorp_batch *batch, void *location, argument 91 return blorp_emit_reloc(batch, locatio 151 emit_urb_config(struct blorp_batch *batch, const struct blorp_params *params) argument 175 blorp_emit_vertex_data(struct blorp_batch *batch, const struct blorp_params *params, struct blorp_address *addr, uint32_t *size) argument 193 blorp_emit_input_varying_data(struct blorp_batch *batch, const struct blorp_params *params, struct blorp_address *addr, uint32_t *size) argument 237 blorp_emit_vertex_buffers(struct blorp_batch *batch, const struct blorp_params *params) argument 285 blorp_emit_vertex_elements(struct blorp_batch *batch, const struct blorp_params *params) argument 412 blorp_emit_vs_config(struct blorp_batch *batch, const struct blorp_params *params) argument 441 blorp_emit_sf_config(struct blorp_batch *batch, const struct blorp_params *params) argument 542 blorp_emit_ps_config(struct blorp_batch *batch, const struct blorp_params *params) argument 788 blorp_emit_depth_stencil_config(struct blorp_batch *batch, const struct blorp_params *params) argument 895 blorp_emit_blend_state(struct blorp_batch *batch, const struct blorp_params *params) argument 938 blorp_emit_color_calc_state(struct blorp_batch *batch, const struct blorp_params *params) argument 967 blorp_emit_depth_stencil_state(struct blorp_batch *batch, const struct blorp_params *params) argument 1042 blorp_emit_surface_state(struct blorp_batch *batch, const struct brw_blorp_surface_info *surface, void *state, uint32_t state_offset, bool is_render_target) argument 1087 blorp_emit_null_surface_state(struct blorp_batch *batch, const struct brw_blorp_surface_info *surface, uint32_t *state) argument 1119 blorp_emit_surface_states(struct blorp_batch *batch, const struct blorp_params *params) argument 1172 blorp_emit_sampler_state(struct blorp_batch *batch, const struct blorp_params *params) argument 1216 blorp_emit_3dstate_multisample(struct blorp_batch *batch, const struct blorp_params *params) argument 1262 blorp_emit_gen8_hiz_op(struct blorp_batch *batch, const struct blorp_params *params) argument 1338 blorp_emit_viewport_state(struct blorp_batch *batch, const struct blorp_params *params) argument 1377 blorp_exec(struct blorp_batch *batch, const struct blorp_params *params) argument [all...] |
/external/tensorflow/tensorflow/core/kernels/ |
H A D | eigen_activations_test.cc | 30 const ptrdiff_t batch = 10; local 34 Tensor<float, 4> input(depth, rows, cols, batch); 37 Tensor<float, 4> result(depth, rows, cols, batch); 40 for (int b = 0; b < batch; ++b) { 54 const ptrdiff_t batch = 10; local 58 Tensor<float, 4> input(depth, rows, cols, batch); 61 Tensor<float, 4> result(depth, rows, cols, batch); 64 for (int b = 0; b < batch; ++b) { 78 const ptrdiff_t batch = 10; local 82 Tensor<float, 4> input(depth, rows, cols, batch); [all...] |
/external/skia/tests/ |
H A D | PathOpsThreadedCommon.cpp | 19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {
|
/external/skqp/tests/ |
H A D | PathOpsThreadedCommon.cpp | 19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {
|