/external/tensorflow/tensorflow/core/kernels/ |
H A D | sparse_reduce_sum_op_test.cc | 53 const auto indices_shape = TensorShape({4, 2}); local 59 AddInputFromArray<int64>(indices_shape, indices); 95 const auto indices_shape = TensorShape({4, 2}); local 101 AddInputFromArray<int64>(indices_shape, indices);
|
H A D | sparse_add_op_test.cc | 63 const auto indices_shape = TensorShape({4, 2}); local 70 AddInputFromArray<int64>(indices_shape, indices); \ 81 Tensor expected_indices(allocator(), DT_INT64, indices_shape); local 103 const auto indices_shape = TensorShape({4, 2}); \ 109 AddInputFromArray<int64>(indices_shape, indices); \ 160 const auto indices_shape = TensorShape({4, 2}); \ 166 auto AddSparseTensor = [indices, indices_shape, shape, \ 168 AddInputFromArray<int64>(indices_shape, indices); \
|
H A D | one_hot_op.cc | 55 const TensorShape& indices_shape = indices.shape(); variable 57 const int indices_dims = indices_shape.dims(); 84 MultiplyWithoutOverflow(indices_shape.num_elements(), depth_v) >= 0, 86 indices_shape.DebugString(), " + [", depth_v, 89 TensorShape output_shape = indices_shape; 104 prefix_dim_size *= indices_shape.dim_size(i); 106 TI suffix_dim_size = indices_shape.num_elements() / prefix_dim_size;
|
H A D | sparse_dense_binary_op_shared_test.cc | 98 const auto indices_shape = TensorShape({4, 2}); local 109 AddInputFromArray<int64>(indices_shape, indices); 127 const auto indices_shape = TensorShape({4, 2}); local 137 AddInputFromArray<int64>(indices_shape, indices); 154 const auto indices_shape = TensorShape({4, 2}); local 164 AddInputFromArray<int64>(indices_shape, indices); 186 const auto indices_shape = TensorShape({4, 2}); local 197 AddInputFromArray<int64>(indices_shape, indices);
|
/external/tensorflow/tensorflow/core/ops/ |
H A D | state_ops.cc | 118 ShapeHandle indices_shape = c->input(1); local 124 TF_RETURN_IF_ERROR(c->Concatenate(indices_shape, var_subshape, &concat));
|
H A D | data_flow_ops.cc | 104 ShapeHandle indices_shape = c->input(i); local 106 if (!c->RankKnown(indices_shape)) { 109 const int64 indices_rank = c->Rank(indices_shape); 111 // Assert that data_shape starts with indices_shape. 114 c->MergePrefix(data_shape, indices_shape, &unused, &unused));
|
H A D | math_ops.cc | 879 ShapeHandle indices_shape; local 880 TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 1, &indices_shape)); 887 TF_RETURN_IF_ERROR(c->Merge(indices_shape, segment_ids_shape, &unused)); 903 ShapeHandle indices_shape; local 904 TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 1, &indices_shape)); 908 TF_RETURN_IF_ERROR(c->Merge(c->input(2), indices_shape, &unused)); 941 ShapeHandle indices_shape; local 942 TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 1, &indices_shape)); 952 TF_RETURN_IF_ERROR(c->Merge(indices_shape, segment_ids_shape, &unused));
|
H A D | array_ops.cc | 894 ShapeHandle indices_shape = c->input(1); 896 TF_RETURN_IF_ERROR(c->Concatenate(indices_shape, params_subshape, &out)); 914 ShapeHandle indices_shape = c->input(1); 922 if (c->RankKnown(params_shape) && c->RankKnown(indices_shape)) { 924 c->Rank(indices_shape) - 1)); 950 c->Concatenate(params_outer_subshape, indices_shape, &out)); 2616 ShapeHandle indices_shape; local 2617 TF_RETURN_IF_ERROR(c->WithRankAtLeast(c->input(0), 1, &indices_shape)); 2624 (c->Value(c->NumElements(indices_shape)) > 0 || 2630 if (c->RankKnown(indices_shape) [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | one_hot_op.cc | 33 const TensorShape indices_shape = ctx->InputShape(0); variable 38 const int indices_dims = indices_shape.dims(); 68 indices_shape, ctx->Input(0), ctx->Input(2),
|
H A D | scatter_nd_op.cc | 32 const TensorShape& indices_shape, 34 if (indices_shape.dims() < 1) { 37 indices_shape.DebugString()); 40 const int64 num_index_dims = indices_shape.dim_size(indices_shape.dims() - 1); 41 const int64 batch_dim = indices_shape.dims() - 1; 48 ", indices.shape: ", indices_shape.DebugString(), 63 if (updates_shape.dim_size(d) != indices_shape.dim_size(d)) { 83 TensorShape indices_shape = context->InputShape(0); variable 96 buffer_shape.num_elements() > 0 || (indices_shape 31 ValidateUpdateShape(const TensorShape& buffer_shape, const TensorShape& indices_shape, const TensorShape& updates_shape) argument [all...] |
H A D | segment_reduction_ops.cc | 47 TensorShape indices_shape = ctx->InputShape(1); variable 52 OP_REQUIRES(ctx, data_shape.dims() >= indices_shape.dims(), 57 for (int d = 0; d < indices_shape.dims(); ++d) { 58 OP_REQUIRES(ctx, (data_shape.dim_size(d) == indices_shape.dim_size(d)), 63 indices_shape.dim_size(d))); 67 buffer_shape.RemoveDimRange(0, indices_shape.dims());
|
H A D | dynamic_stitch_op.cc | 70 TensorShape indices_shape; variable 73 &indices_shape)); 76 ctx, TensorShapeUtils::StartsWith(data_shape, indices_shape), 80 "].shape = ", indices_shape.DebugString())); 84 data_shape, indices_shape), 87 input_num, "].shape[", indices_shape.dims(), 92 "].shape = ", indices_shape.DebugString())); 96 {indices_shape.num_elements()},
|
H A D | variable_ops.cc | 118 auto indices_shape = ctx->InputShape(1); variable 122 ctx, XlaGather(resource_handle, resource_shape, indices, indices_shape,
|
H A D | gather_op.cc | 32 TensorShape indices_shape, int64 axis, bool indices_are_nd, 40 CHECK_GE(indices_shape.dims(), 1); 41 num_index_dims = indices_shape.dim_size(indices_shape.dims() - 1); 42 indices_shape.RemoveLastDims(1); 50 const int64 num_indices = indices_shape.num_elements(); 73 out_shape.AppendShape(indices_shape); 173 auto indices_shape = context->InputShape(1); variable 200 context, XlaGather(input, input_shape, indices, indices_shape, axis, 222 TensorShape indices_shape variable 29 XlaGather(const xla::ComputationDataHandle& input, const TensorShape& input_shape, const xla::ComputationDataHandle& indices, TensorShape indices_shape, int64 axis, bool indices_are_nd, DataType dtype, DataType index_type, xla::ComputationBuilder* builder, xla::ComputationDataHandle* gather_output) argument [all...] |
H A D | tensor_array_ops.cc | 306 const TensorShape indices_shape = ctx->InputShape(1); variable 307 OP_REQUIRES(ctx, indices_shape.dims() == 1, 343 XlaGather(ta, ta_shape, indices, indices_shape, /*axis=*/0, 374 const TensorShape indices_shape = ctx->InputShape(1); variable 375 OP_REQUIRES(ctx, indices_shape.dims() >= 1, 377 const int num_indices = indices_shape.dim_size(0);
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
H A D | embedding_lookup_sparse_test.cc | 36 std::initializer_list<int> indices_shape, 48 BuildInterpreter({lookup_shape, indices_shape, dense_shape_shape, 34 EmbeddingLookupSparseOpModel(CombinerType type, std::initializer_list<int> lookup_shape, std::initializer_list<int> indices_shape, std::initializer_list<int> dense_shape_shape, std::initializer_list<int> value_shape) argument
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
H A D | xla_helpers.cc | 228 const TensorShape& indices_shape, 233 const int indices_dims = indices_shape.dims(); 236 TensorShape output_shape = indices_shape; 263 std::vector<int64> broadcast_dims(indices_shape.dims()); 226 OneHot(xla::ComputationBuilder* builder, int64 depth, int axis, DataType index_type, const TensorShape& indices_shape, const xla::ComputationDataHandle& indices, const xla::ComputationDataHandle& on_value, const xla::ComputationDataHandle& off_value, xla::ComputationDataHandle* one_hot) argument
|
/external/tensorflow/tensorflow/core/util/ |
H A D | example_proto_helper.h | 85 TensorShape indices_shape; member in struct:tensorflow::VarLenFeatureBatchShapes
|
H A D | example_proto_helper.cc | 310 output_shapes->indices_shape.AddDim(total_num_features); 311 output_shapes->indices_shape.AddDim(2); 378 const TensorShape& indices_shape = sparse_tensor_batch_shapes.indices_shape; local 383 Tensor(allocator, DT_INT64, indices_shape);
|
/external/tensorflow/tensorflow/contrib/lite/toco/graph_transformations/ |
H A D | propagate_fixed_sizes.cc | 968 const auto& indices_shape = indices_array.shape(); local 973 QCHECK_EQ(indices_shape.dimensions_count(), 1); 976 // where the dimension of indices_shape is used. 979 output_dims->push_back(indices_shape.dims(0));
|
/external/tensorflow/tensorflow/core/framework/ |
H A D | common_shape_fns.cc | 1286 Status ValidateSparseTensor(InferenceContext* c, ShapeHandle indices_shape, argument 1290 TF_RETURN_IF_ERROR(c->WithRank(indices_shape, 2, &unused_shape)); 1295 DimensionHandle num_index_elements_dim = c->Dim(indices_shape, 0); 1310 DimensionHandle index_rank_dim = c->Dim(indices_shape, 1); 1332 ShapeHandle indices_shape; local 1333 TF_RETURN_IF_ERROR(c->WithRankAtLeast(c->input(1), 1, &indices_shape)); 1338 (c->Value(c->NumElements(indices_shape)) > 0 || 1344 if (c->RankKnown(indices_shape) && c->RankKnown(updates_shape)) { 1345 const int64 num_outer_dims = c->Rank(indices_shape) - 1; 1346 const DimensionHandle index_size = c->Dim(indices_shape, [all...] |