/external/tensorflow/tensorflow/python/training/ |
H A D | input_test.py | 759 self.assertIs(None, sparse.dense_shape.get_shape().num_elements()) 761 self.assertIs(None, batched.dense_shape.get_shape().num_elements()) 768 self.assertIs(None, sparse.dense_shape.get_shape().num_elements()) 770 self.assertIs(None, batched.dense_shape.get_shape().num_elements()) 882 self.assertIs(None, sparse.dense_shape.get_shape().num_elements()) 884 self.assertIs(None, batched.dense_shape.get_shape().num_elements()) 891 self.assertIs(None, sparse.dense_shape.get_shape().num_elements()) 894 self.assertIs(None, batched.dense_shape.get_shape().num_elements()) 901 self.assertIs(None, sparse.dense_shape.get_shape().num_elements()) 904 self.assertIs(None, batched.dense_shape.get_shape().num_elements()) [all...] |
/external/mesa3d/src/gallium/state_trackers/va/ |
H A D | picture_hevc.c | 35 assert(buf->size >= sizeof(VAPictureParameterBufferHEVC) && buf->num_elements == 1); 183 assert(buf->size >= sizeof(VAIQMatrixBufferH264) && buf->num_elements == 1); 196 assert(buf->size >= sizeof(VASliceParameterBufferHEVC) && buf->num_elements == 1);
|
/external/tensorflow/tensorflow/core/framework/ |
H A D | tensor_shape.h | 69 int64 num_elements() const { return num_elements_; } function in class:tensorflow::TensorShapeRep 248 bool IsFullyDefined() const { return !kIsPartial || num_elements() != -1; } 387 static Status NumElements(gtl::ArraySlice<int64> shape, int64* num_elements);
|
/external/tensorflow/tensorflow/core/kernels/ |
H A D | maxpooling_op.cc | 261 ksize.resize(tensor_ksize.shape().num_elements()); 266 stride.resize(tensor_stride.shape().num_elements()); 385 ksize.resize(tensor_ksize.shape().num_elements()); 390 stride.resize(tensor_stride.shape().num_elements()); 490 ksize.resize(tensor_ksize.shape().num_elements()); 495 stride.resize(tensor_stride.shape().num_elements()); 691 ksize.resize(tensor_ksize.shape().num_elements()); 696 stride.resize(tensor_stride.shape().num_elements()); 830 ksize.resize(tensor_ksize.shape().num_elements()); 835 stride.resize(tensor_stride.shape().num_elements()); [all...] |
H A D | padding_fifo_queue.cc | 69 void PaddingFIFOQueue::TryDequeueMany(int num_elements, OpKernelContext* ctx, argument 72 if (num_elements == 0) { 99 num_elements, [callback]() { callback(Tuple()); }, ctx, cm, token,
|
H A D | priority_queue.cc | 238 void PriorityQueue::TryDequeueMany(int num_elements, OpKernelContext* ctx, argument 248 if (num_elements == 0) { 300 num_elements, [callback]() { callback(Tuple()); }, ctx, cm, token,
|
H A D | extract_image_patches_op.cc | 103 if (out_shape.num_elements() == 0) {
|
H A D | histogram_op.cc | 86 OP_REQUIRES(ctx, (value_range_tensor.shape().num_elements() == 2),
|
H A D | list_kernels.cc | 268 c, c->input(0).shape().num_elements() == 1, 358 int32 num_elements = c->input(1).scalar<int32>()(); variable 362 output.tensors.resize(num_elements, Tensor(DT_INVALID)); 382 .HostMemory("num_elements"), 395 c, c->input(0).shape().num_elements() == 1,
|
H A D | mkl_conv_ops.h | 386 if (input_tf_shape.num_elements() == 0 || 387 filter_tf_shape.num_elements() == 0 || 388 outbprop_tf_shape.num_elements() == 0) { 399 for (size_t i = 0; i < output_tf_shape.num_elements(); ++i) {
|
H A D | random_shuffle_queue_op.cc | 57 void TryDequeueMany(int num_elements, OpKernelContext* ctx, 274 void RandomShuffleQueue::TryDequeueMany(int num_elements, OpKernelContext* ctx, argument 284 if (num_elements == 0) { 336 num_elements, [callback]() { callback(Tuple()); }, ctx, cm, token,
|
H A D | stateless_random_ops.cc | 50 if (shape.num_elements() == 0) return;
|
H A D | debug_ops.h | 236 // Use DT_INT64/int64 to be consistent with TensorShape::num_elements(). 245 for (int64 i = 0; i < input_shape.num_elements(); ++i) { 301 element_count = input_shape.num_elements();
|
H A D | tensor_array_ops.cc | 646 if (output_shape.num_elements() == 0) { 653 output_tensor->shaped<T, 2>({1, output_shape.num_elements()}); 899 if (output_shape.num_elements() > 0) { 901 output_tensor->shaped<T, 2>({1, output_shape.num_elements()}); 1087 {1, num_values, element_shape.num_elements()}); 1091 element_shape.num_elements()}; 1103 tensor_value_i->shaped<T, 3>({1, 1, element_shape.num_elements()}); 1106 if (element_shape.num_elements() > 0) {
|
/external/tensorflow/tensorflow/tools/graph_transforms/ |
H A D | summarize_graph_main.cc | 218 const size_t num_elements = tensor.NumElements(); local 220 const_parameter_count += num_elements; 222 variable_parameter_count += num_elements;
|
/external/mesa3d/src/gallium/drivers/etnaviv/ |
H A D | etnaviv_internal.h | 199 unsigned num_elements; member in struct:compiled_vertex_elements_state
|
/external/mesa3d/src/gallium/drivers/radeonsi/ |
H A D | si_debug.c | 368 unsigned num_elements, 381 for (i = 0; i < num_elements; i++) { 454 unsigned num_elements[] = { local 468 num_elements[i], f); 365 si_dump_descriptor_list(struct si_descriptors *desc, const char *shader_name, const char *elem_name, unsigned num_elements, FILE *f) argument
|
/external/mesa3d/src/gallium/drivers/vc4/ |
H A D | vc4_draw.c | 151 uint32_t num_elements_emit = MAX2(vtx->num_elements, 1); 185 for (int i = 0; i < vtx->num_elements; i++) { 211 if (vtx->num_elements == 0) { 226 assert(vtx->num_elements <= 8);
|
H A D | vc4_state.c | 354 vc4_vertex_state_create(struct pipe_context *pctx, unsigned num_elements, argument 362 memcpy(so->pipe, elements, sizeof(*elements) * num_elements); 363 so->num_elements = num_elements;
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | bcast_ops.cc | 100 for (int64 i = 0; i < in_shape.num_elements(); ++i) {
|
H A D | diag_op.cc | 108 int64 size = input_shape.num_elements(); 174 int64 size = input_shape.num_elements();
|
/external/tensorflow/tensorflow/tools/benchmark/ |
H A D | benchmark_model.cc | 203 int64 output_count = output_shape.num_elements(); 216 int64 output_count = output_shape.num_elements(); 223 int64 output_count = output_shape.num_elements(); 605 int64 total_size = inputs[0].shape.num_elements();
|
/external/webp/include/webp/ |
H A D | mux.h | 369 // num_elements - (out) number of chunks with the given chunk id 371 // WEBP_MUX_INVALID_ARGUMENT - if mux, or num_elements is NULL. 374 WebPChunkId id, int* num_elements);
|
/external/webp/src/mux/ |
H A D | muxread.c | 523 WebPChunkId id, int* num_elements) { 524 if (mux == NULL || num_elements == NULL) { 529 *num_elements = MuxImageCount(mux->images_, id); 533 *num_elements = CountChunks(*chunk_list, kChunks[idx].tag); 522 WebPMuxNumChunks(const WebPMux* mux, WebPChunkId id, int* num_elements) argument
|
/external/webp/src/webp/ |
H A D | mux.h | 369 // num_elements - (out) number of chunks with the given chunk id 371 // WEBP_MUX_INVALID_ARGUMENT - if mux, or num_elements is NULL. 374 WebPChunkId id, int* num_elements);
|