Searched refs:batch (Results 1 - 25 of 412) sorted by relevance

1234567891011>>

/external/mesa3d/src/gallium/drivers/freedreno/
H A Dfreedreno_batch.c38 batch_init(struct fd_batch *batch) argument
40 struct fd_context *ctx = batch->ctx;
44 util_queue_fence_init(&batch->flush_fence);
56 batch->draw = fd_ringbuffer_new(ctx->screen->pipe, size);
57 batch->binning = fd_ringbuffer_new(ctx->screen->pipe, size);
58 batch->gmem = fd_ringbuffer_new(ctx->screen->pipe, size);
60 fd_ringbuffer_set_parent(batch->gmem, NULL);
61 fd_ringbuffer_set_parent(batch->draw, batch->gmem);
62 fd_ringbuffer_set_parent(batch
92 struct fd_batch *batch = CALLOC_STRUCT(fd_batch); local
111 batch_fini(struct fd_batch *batch) argument
139 batch_flush_reset_dependencies(struct fd_batch *batch, bool flush) argument
154 batch_reset_resources_locked(struct fd_batch *batch) argument
171 batch_reset_resources(struct fd_batch *batch) argument
179 batch_reset(struct fd_batch *batch) argument
193 fd_batch_reset(struct fd_batch *batch) argument
200 __fd_batch_destroy(struct fd_batch *batch) argument
223 __fd_batch_describe(char* buf, const struct fd_batch *batch) argument
229 fd_batch_sync(struct fd_batch *batch) argument
239 struct fd_batch *batch = job; local
248 struct fd_batch *batch = job; local
253 batch_flush(struct fd_batch *batch) argument
298 fd_batch_flush(struct fd_batch *batch, bool sync) argument
314 batch_depends_on(struct fd_batch *batch, struct fd_batch *other) argument
330 batch_add_dep(struct fd_batch *batch, struct fd_batch *dep) argument
352 fd_batch_resource_used(struct fd_batch *batch, struct fd_resource *rsc, bool write) argument
400 fd_batch_check_size(struct fd_batch *batch) argument
415 fd_wfi(struct fd_batch *batch, struct fd_ringbuffer *ring) argument
[all...]
H A Dfreedreno_batch.h62 /* A batch tracks everything about a cmdstream batch/submit, including the
119 unsigned num_draws; /* number of draws in current batch */
122 * batch. Used at the tile rendering step (fd_gmem_render_tiles(),
161 /* next sample offset.. incremented for each sample in the batch/
171 /* which sample providers were active in the current batch: */
179 /* list of samples in current batch: */
188 /* Set of resources used by currently-unsubmitted batch (read or
193 /** key in batch-cache (if not null): */
203 void fd_batch_reset(struct fd_batch *batch);
225 fd_batch_reference(struct fd_batch **ptr, struct fd_batch *batch) argument
240 fd_batch_reference_locked(struct fd_batch **ptr, struct fd_batch *batch) argument
262 fd_reset_wfi(struct fd_batch *batch) argument
272 fd_event_write(struct fd_batch *batch, struct fd_ringbuffer *ring, enum vgt_event_type evt) argument
[all...]
H A Dfreedreno_gmem.c107 calculate_tiles(struct fd_batch *batch) argument
109 struct fd_context *ctx = batch->ctx;
111 struct pipe_scissor_state *scissor = &batch->max_scissor;
112 struct pipe_framebuffer_state *pfb = &batch->framebuffer;
123 bool has_zs = !!(batch->resolve & (FD_BUFFER_DEPTH | FD_BUFFER_STENCIL));
310 render_tiles(struct fd_batch *batch) argument
312 struct fd_context *ctx = batch->ctx;
316 ctx->emit_tile_init(batch);
318 if (batch->restore)
327 ctx->emit_tile_prep(batch, til
350 render_sysmem(struct fd_batch *batch) argument
367 flush_ring(struct fd_batch *batch) argument
381 fd_gmem_render_tiles(struct fd_batch *batch) argument
427 fd_gmem_render_noop(struct fd_batch *batch) argument
458 fd_gmem_needs_restore(struct fd_batch *batch, struct fd_tile *tile, uint32_t buffers) argument
[all...]
H A Dfreedreno_query_hw.c64 get_sample(struct fd_batch *batch, struct fd_ringbuffer *ring, argument
67 struct fd_context *ctx = batch->ctx;
73 if (!batch->sample_cache[idx]) {
75 ctx->sample_providers[idx]->get_sample(batch, ring);
76 fd_hw_sample_reference(ctx, &batch->sample_cache[idx], new_samp);
77 util_dynarray_append(&batch->samples, struct fd_hw_sample *, new_samp);
78 batch->needs_flush = true;
81 fd_hw_sample_reference(ctx, &samp, batch->sample_cache[idx]);
87 clear_sample_cache(struct fd_batch *batch) argument
91 for (i = 0; i < ARRAY_SIZE(batch
103 resume_query(struct fd_batch *batch, struct fd_hw_query *hq, struct fd_ringbuffer *ring) argument
119 pause_query(struct fd_batch *batch, struct fd_hw_query *hq, struct fd_ringbuffer *ring) argument
160 struct fd_batch *batch = ctx->batch; local
186 struct fd_batch *batch = ctx->batch; local
340 fd_hw_sample_init(struct fd_batch *batch, uint32_t size) argument
386 fd_hw_query_prepare(struct fd_batch *batch, uint32_t num_tiles) argument
408 fd_hw_query_prepare_tile(struct fd_batch *batch, uint32_t n, struct fd_ringbuffer *ring) argument
424 fd_hw_query_set_stage(struct fd_batch *batch, struct fd_ringbuffer *ring, enum fd_render_stage stage) argument
458 fd_hw_query_enable(struct fd_batch *batch, struct fd_ringbuffer *ring) argument
[all...]
H A Dfreedreno_draw.c43 resource_read(struct fd_batch *batch, struct pipe_resource *prsc) argument
47 fd_batch_resource_used(batch, fd_resource(prsc), false);
51 resource_written(struct fd_batch *batch, struct pipe_resource *prsc) argument
55 fd_batch_resource_used(batch, fd_resource(prsc), true);
62 struct fd_batch *batch = ctx->batch; local
63 struct pipe_framebuffer_state *pfb = &batch->framebuffer;
88 fd_batch_reset(batch);
92 batch->blit = ctx->in_blit;
93 batch
284 struct fd_batch *batch = ctx->batch; local
[all...]
H A Dfreedreno_batch_cache.c39 * The batch cache provides lookup for mapping pipe_framebuffer_state
40 * to a batch.
54 * Each batch needs to hold a reference to each resource it depends on (ie.
59 * When a resource is destroyed, we need to remove entries in the batch
64 * When a batch has weak reference to no more resources (ie. all the
65 * surfaces it rendered to are destroyed) the batch can be destroyed.
68 * surfaces are destroyed before the batch is submitted.
70 * If (for example), batch writes to zsbuf but that surface is destroyed
71 * before batch is submitted, we can skip gmem2mem (but still need to
72 * alloc gmem space as before. If the batch depende
136 struct fd_batch *batch = NULL; local
159 struct fd_batch *batch; local
172 fd_bc_invalidate_batch(struct fd_batch *batch, bool destroy) argument
208 struct fd_batch *batch; local
233 struct fd_batch *batch; local
315 struct fd_batch *batch = NULL; local
[all...]
H A Dfreedreno_query_hw.h80 * at least once before the first ->get_sample() in a batch.
87 struct fd_hw_sample *(*get_sample)(struct fd_batch *batch,
106 * the sample is constructed, but when the batch is submitted.
130 struct list_head list; /* list-node in batch->active_queries */
143 struct fd_hw_sample * fd_hw_sample_init(struct fd_batch *batch, uint32_t size);
146 void fd_hw_query_prepare(struct fd_batch *batch, uint32_t num_tiles);
147 void fd_hw_query_prepare_tile(struct fd_batch *batch, uint32_t n,
149 void fd_hw_query_set_stage(struct fd_batch *batch,
151 void fd_hw_query_enable(struct fd_batch *batch, struct fd_ringbuffer *ring);
/external/mesa3d/src/gallium/drivers/i915/
H A Di915_batchbuffer.h37 i915_winsys_batchbuffer_space(struct i915_winsys_batchbuffer *batch) argument
39 return batch->size - (batch->ptr - batch->map);
43 i915_winsys_batchbuffer_check(struct i915_winsys_batchbuffer *batch, argument
46 return dwords * 4 <= i915_winsys_batchbuffer_space(batch);
50 i915_winsys_batchbuffer_dword_unchecked(struct i915_winsys_batchbuffer *batch, argument
53 *(unsigned *)batch->ptr = dword;
54 batch->ptr += 4;
58 i915_winsys_batchbuffer_float(struct i915_winsys_batchbuffer *batch, argument
68 i915_winsys_batchbuffer_dword(struct i915_winsys_batchbuffer *batch, unsigned dword) argument
76 i915_winsys_batchbuffer_write(struct i915_winsys_batchbuffer *batch, void *data, size_t size) argument
87 i915_winsys_validate_buffers(struct i915_winsys_batchbuffer *batch, struct i915_winsys_buffer **buffers, int num_of_buffers) argument
95 i915_winsys_batchbuffer_reloc(struct i915_winsys_batchbuffer *batch, struct i915_winsys_buffer *buffer, enum i915_winsys_buffer_usage usage, size_t offset, boolean fenced) argument
[all...]
H A Di915_batch.h36 (i915_winsys_batchbuffer_check(i915->batch, dwords))
39 i915_winsys_batchbuffer_dword(i915->batch, dword)
42 i915_winsys_batchbuffer_float(i915->batch, f)
45 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, false)
48 i915_winsys_batchbuffer_reloc(i915->batch, buf, usage, offset, true)
H A Di915_flush.c48 if (!i915->batch)
54 if (!fence && (i915->batch->map == i915->batch->ptr)) {
79 struct i915_winsys_batchbuffer *batch = i915->batch; local
81 batch->iws->batchbuffer_flush(batch, fence, flags);
/external/tensorflow/tensorflow/core/kernels/batching_util/
H A Dbatch_scheduler_test.cc41 Batch<FakeTask> batch; local
43 EXPECT_EQ(0, batch.num_tasks());
44 EXPECT_TRUE(batch.empty());
45 EXPECT_EQ(0, batch.size());
46 EXPECT_FALSE(batch.IsClosed());
49 batch.AddTask(std::unique_ptr<FakeTask>(task0));
51 EXPECT_EQ(1, batch.num_tasks());
52 EXPECT_FALSE(batch.empty());
53 EXPECT_EQ(task0->size(), batch.size());
54 EXPECT_EQ(task0->size(), batch
84 Batch<FakeTask> batch; local
98 Batch<FakeTask>* batch = new Batch<FakeTask>; local
[all...]
/external/mesa3d/src/mesa/drivers/dri/i965/
H A Dintel_batchbuffer.h14 * Number of bytes to reserve for commands necessary to complete a batch.
18 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes)
42 void intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr,
44 void intel_batchbuffer_free(struct intel_batchbuffer *batch);
66 uint32_t intel_batchbuffer_reloc(struct intel_batchbuffer *batch,
72 uint64_t intel_batchbuffer_reloc64(struct intel_batchbuffer *batch,
79 #define USED_BATCH(batch) ((uintptr_t)((batch).map_next - (batch).map))
98 intel_batchbuffer_space(struct intel_batchbuffer *batch) argument
106 intel_batchbuffer_emit_dword(struct intel_batchbuffer *batch, GLuint dword) argument
116 intel_batchbuffer_emit_float(struct intel_batchbuffer *batch, float f) argument
136 struct intel_batchbuffer *batch = &brw->batch; local
[all...]
H A DgenX_blorp_exec.c38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n) argument
40 assert(batch->blorp->driver_ctx == batch->driver_batch);
41 struct brw_context *brw = batch->driver_batch;
44 uint32_t *map = brw->batch.map_next;
45 brw->batch.map_next += n;
51 blorp_emit_reloc(struct blorp_batch *batch, argument
54 assert(batch->blorp->driver_ctx == batch->driver_batch);
55 struct brw_context *brw = batch
72 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, struct blorp_address address, uint32_t delta) argument
93 blorp_alloc_dynamic_state(struct blorp_batch *batch, enum aub_state_struct_type type, uint32_t size, uint32_t alignment, uint32_t *offset) argument
106 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries, unsigned state_size, unsigned state_alignment, uint32_t *bt_offset, uint32_t *surface_offsets, void **surface_maps) argument
127 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size, struct blorp_address *addr) argument
160 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size) argument
168 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size) argument
187 blorp_exec(struct blorp_batch *batch, const struct blorp_params *params) argument
[all...]
H A Dintel_batchbuffer.c39 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr,
43 intel_batchbuffer_init(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, argument
46 intel_batchbuffer_reset(batch, bufmgr, has_llc);
49 batch->cpu_map = malloc(BATCH_SZ);
50 batch->map = batch->cpu_map;
51 batch->map_next = batch->cpu_map;
56 intel_batchbuffer_reset(struct intel_batchbuffer *batch, dri_bufmgr *bufmgr, argument
59 if (batch
109 intel_batchbuffer_free(struct intel_batchbuffer *batch) argument
146 struct intel_batchbuffer *batch = &brw->batch; local
324 struct intel_batchbuffer *batch = &brw->batch; local
442 intel_batchbuffer_reloc(struct intel_batchbuffer *batch, drm_intel_bo *buffer, uint32_t offset, uint32_t read_domains, uint32_t write_domain, uint32_t delta) argument
463 intel_batchbuffer_reloc64(struct intel_batchbuffer *batch, drm_intel_bo *buffer, uint32_t offset, uint32_t read_domains, uint32_t write_domain, uint32_t delta) argument
[all...]
H A Dbrw_state_batch.c44 struct intel_batchbuffer *batch = &brw->batch; local
52 batch->bo->size / 32);
76 * Generate a set of aub file annotations for the current batch buffer, and
79 * The "used" section of the batch buffer (the portion containing batch
81 * batch buffer (which contains data structures pointed to by batch commands)
91 4 * USED_BATCH(brw->batch));
102 drm_intel_bufmgr_gem_set_aub_annotations(brw->batch
129 struct intel_batchbuffer *batch = &brw->batch; local
[all...]
H A Dbrw_blorp.h74 void gen6_blorp_exec(struct blorp_batch *batch,
76 void gen7_blorp_exec(struct blorp_batch *batch,
78 void gen75_blorp_exec(struct blorp_batch *batch,
80 void gen8_blorp_exec(struct blorp_batch *batch,
82 void gen9_blorp_exec(struct blorp_batch *batch,
/external/mesa3d/src/gallium/winsys/i915/drm/
H A Di915_drm_batchbuffer.c30 i915_drm_batchbuffer(struct i915_winsys_batchbuffer *batch) argument
32 return (struct i915_drm_batchbuffer *)batch;
36 i915_drm_batchbuffer_reset(struct i915_drm_batchbuffer *batch) argument
38 struct i915_drm_winsys *idws = i915_drm_winsys(batch->base.iws);
40 if (batch->bo)
41 drm_intel_bo_unreference(batch->bo);
42 batch->bo = drm_intel_bo_alloc(idws->gem_manager,
44 batch->actual_size,
47 memset(batch->base.map, 0, batch
57 struct i915_drm_batchbuffer *batch = CALLOC_STRUCT(i915_drm_batchbuffer); local
75 i915_drm_batchbuffer_validate_buffers(struct i915_winsys_batchbuffer *batch, struct i915_winsys_buffer **buffer, int num_of_buffers) argument
100 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local
165 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local
221 struct i915_drm_batchbuffer *batch = i915_drm_batchbuffer(ibatch); local
[all...]
/external/mesa3d/src/mesa/drivers/dri/i915/
H A Dintel_batchbuffer.c43 intel->batch.cpu_map = malloc(intel->maxBatchSize);
44 intel->batch.map = intel->batch.cpu_map;
50 if (intel->batch.last_bo != NULL) {
51 drm_intel_bo_unreference(intel->batch.last_bo);
52 intel->batch.last_bo = NULL;
54 intel->batch.last_bo = intel->batch.bo;
56 intel->batch.bo = drm_intel_bo_alloc(intel->bufmgr, "batchbuffer",
59 intel->batch
75 struct intel_batchbuffer *batch = &intel->batch; local
116 struct intel_batchbuffer *batch = &intel->batch; local
[all...]
H A Dintel_batchbuffer.h15 * Number of bytes to reserve for commands necessary to complete a batch.
19 * - Optional MI_NOOP for ensuring the batch length is qword aligned (4 bytes)
76 return (intel->batch.bo->size - intel->batch.reserved_space)
77 - intel->batch.used*4;
87 intel->batch.map[intel->batch.used++] = dword;
112 intel->batch.emit = intel->batch.used;
114 intel->batch
122 struct intel_batchbuffer *batch = &intel->batch; local
[all...]
/external/mesa3d/src/intel/vulkan/
H A DgenX_blorp_exec.c38 blorp_emit_dwords(struct blorp_batch *batch, unsigned n) argument
40 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
41 return anv_batch_emit_dwords(&cmd_buffer->batch, n);
45 blorp_emit_reloc(struct blorp_batch *batch, argument
48 struct anv_cmd_buffer *cmd_buffer = batch->driver_batch;
49 assert(cmd_buffer->batch.start <= location &&
50 location < cmd_buffer->batch.end);
51 return anv_batch_emit_reloc(&cmd_buffer->batch, location,
56 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset, argument
59 struct anv_cmd_buffer *cmd_buffer = batch
65 blorp_alloc_dynamic_state(struct blorp_batch *batch, enum aub_state_struct_type type, uint32_t size, uint32_t alignment, uint32_t *offset) argument
81 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries, unsigned state_size, unsigned state_alignment, uint32_t *bt_offset, uint32_t *surface_offsets, void **surface_maps) argument
109 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size, struct blorp_address *addr) argument
138 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size) argument
146 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size) argument
164 blorp_exec(struct blorp_batch *batch, const struct blorp_params *params) argument
[all...]
/external/jsoncpp/src/lib_json/
H A Djson_batchallocator.h18 * This memory allocator allocates memory for a batch of object (specified by
44 for (BatchInfo* batch = batches_; batch;) {
45 BatchInfo* nextBatch = batch->next_;
46 free(batch);
47 batch = nextBatch;
66 if (!currentBatch_) // no free batch found, allocate a new one
103 BatchInfo* batch = static_cast<BatchInfo*>(malloc(mallocSize)); local
104 batch->next_ = 0;
105 batch
[all...]
/external/mesa3d/src/intel/blorp/
H A Dblorp_genX_exec.h48 blorp_emit_dwords(struct blorp_batch *batch, unsigned n);
51 blorp_emit_reloc(struct blorp_batch *batch,
55 blorp_alloc_dynamic_state(struct blorp_batch *batch,
61 blorp_alloc_vertex_buffer(struct blorp_batch *batch, uint32_t size,
65 blorp_alloc_binding_table(struct blorp_batch *batch, unsigned num_entries,
71 blorp_flush_range(struct blorp_batch *batch, void *start, size_t size);
74 blorp_surface_reloc(struct blorp_batch *batch, uint32_t ss_offset,
78 blorp_emit_urb_config(struct blorp_batch *batch, unsigned vs_entry_size);
85 _blorp_combine_address(struct blorp_batch *batch, void *location, argument
91 return blorp_emit_reloc(batch, locatio
151 emit_urb_config(struct blorp_batch *batch, const struct blorp_params *params) argument
175 blorp_emit_vertex_data(struct blorp_batch *batch, const struct blorp_params *params, struct blorp_address *addr, uint32_t *size) argument
193 blorp_emit_input_varying_data(struct blorp_batch *batch, const struct blorp_params *params, struct blorp_address *addr, uint32_t *size) argument
237 blorp_emit_vertex_buffers(struct blorp_batch *batch, const struct blorp_params *params) argument
285 blorp_emit_vertex_elements(struct blorp_batch *batch, const struct blorp_params *params) argument
412 blorp_emit_vs_config(struct blorp_batch *batch, const struct blorp_params *params) argument
441 blorp_emit_sf_config(struct blorp_batch *batch, const struct blorp_params *params) argument
542 blorp_emit_ps_config(struct blorp_batch *batch, const struct blorp_params *params) argument
788 blorp_emit_depth_stencil_config(struct blorp_batch *batch, const struct blorp_params *params) argument
895 blorp_emit_blend_state(struct blorp_batch *batch, const struct blorp_params *params) argument
938 blorp_emit_color_calc_state(struct blorp_batch *batch, const struct blorp_params *params) argument
967 blorp_emit_depth_stencil_state(struct blorp_batch *batch, const struct blorp_params *params) argument
1042 blorp_emit_surface_state(struct blorp_batch *batch, const struct brw_blorp_surface_info *surface, void *state, uint32_t state_offset, bool is_render_target) argument
1087 blorp_emit_null_surface_state(struct blorp_batch *batch, const struct brw_blorp_surface_info *surface, uint32_t *state) argument
1119 blorp_emit_surface_states(struct blorp_batch *batch, const struct blorp_params *params) argument
1172 blorp_emit_sampler_state(struct blorp_batch *batch, const struct blorp_params *params) argument
1216 blorp_emit_3dstate_multisample(struct blorp_batch *batch, const struct blorp_params *params) argument
1262 blorp_emit_gen8_hiz_op(struct blorp_batch *batch, const struct blorp_params *params) argument
1338 blorp_emit_viewport_state(struct blorp_batch *batch, const struct blorp_params *params) argument
1377 blorp_exec(struct blorp_batch *batch, const struct blorp_params *params) argument
[all...]
/external/tensorflow/tensorflow/core/kernels/
H A Deigen_activations_test.cc30 const ptrdiff_t batch = 10; local
34 Tensor<float, 4> input(depth, rows, cols, batch);
37 Tensor<float, 4> result(depth, rows, cols, batch);
40 for (int b = 0; b < batch; ++b) {
54 const ptrdiff_t batch = 10; local
58 Tensor<float, 4> input(depth, rows, cols, batch);
61 Tensor<float, 4> result(depth, rows, cols, batch);
64 for (int b = 0; b < batch; ++b) {
78 const ptrdiff_t batch = 10; local
82 Tensor<float, 4> input(depth, rows, cols, batch);
[all...]
/external/skia/tests/
H A DPathOpsThreadedCommon.cpp19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {
/external/skqp/tests/
H A DPathOpsThreadedCommon.cpp19 SkTaskGroup().batch(fRunnables.count(), [&](int i) {

Completed in 1072 milliseconds

1234567891011>>