/external/tensorflow/tensorflow/core/kernels/ |
H A D | split_op_test.cc | 29 static Graph* MakeGraph(int split_dim, int num_split, argument 33 in_shape.set_dim(split_dim, in_shape.dim_size(split_dim) * num_split); 41 .Attr("num_split", num_split) 46 #define BM_SPLIT_1D(num_split, chunk_size) \ 47 static void BM_Split_1d_##num_split##_##chunk_size(int iters) { \ 49 testing::ItemsProcessed(static_cast<int64>(iters) * num_split * \ 52 strings::Printf("1-D %d chunks of %d each", num_split, chunk_size); \ 55 auto g = MakeGraph(/* split_dim = */ 0, num_split, {chunk_size}); \ 59 BENCHMARK(BM_Split_1d_##num_split## [all...] |
H A D | split_v_op_test.cc | 70 .Attr("num_split", static_cast<int64>(size_splits.size())) 75 #define BM_SPLITV_1D(num_split, total_size) \ 76 static void BM_SplitV_1d_##num_split##_##total_size(int iters) { \ 80 strings::Printf("1-D %d chunks totaling %d", num_split, total_size); \ 84 GenerateRandomIntsWithSum(total_size, num_split), \ 89 BENCHMARK(BM_SplitV_1d_##num_split##_##total_size); 91 #define BM_SPLITV_2D(split_dim, num_split, total_size0, total_size1) \ 93 BM_SplitV_2d_##split_dim##_##num_split##_##total_size0##_##total_size1( \ 101 num_split, split_dim, total_size0, total_size1); \ 106 GenerateRandomIntsWithSum(total_size_vec[split_dim], num_split), \ [all...] |
H A D | split_op.cc | 55 const int32 num_split = num_outputs(); 64 context, num_split > 0, 66 "Number of ways to split should be > 0, but got ", num_split)); 68 OP_REQUIRES(context, input_shape.dim_size(split_dim) % num_split == 0, 73 ") ", "and num_split ", num_split)); 74 // Special case 1: num_split == 1. Nothing to do. 75 if (num_split == 1) { 92 const int64 delta = input_shape.dim_size(0) / num_split; 93 for (int i = 0; i < num_split; [all...] |
H A D | split_v_op.cc | 54 const int32 num_split = context->num_outputs(); local 65 split_tensor.dims() == 1 && split_tensor.NumElements() == num_split, 79 context, num_split > 0, 81 "Number of ways to split should be > 0, but got ", num_split)); 91 // Special case 1: num_split == 1. Nothing to do. 92 if (num_split == 1) { 145 for (int i = 0; i < num_split; ++i) { 191 const int32 num_split = Base::num_outputs(); variable 216 std::vector<int64> split_start_points(num_split); 217 for (int i = 0; i < num_split; 303 const int32 num_split = Base::num_outputs(); variable [all...] |
H A D | split_lib_gpu.cu.cc | 75 const int32 num_split = output_ptr_data.size; local 80 eigen_assert(split_dim_size % num_split == 0); 83 int32 piece_size = split_dim_size / num_split; 170 const int32 num_split = output_ptr_data.size; local 177 int32 piece_size = suffix_dim_size / num_split;
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | split_op.cc | 36 const int32 num_split = num_outputs(); variable 66 ctx, num_split > 0, 68 "Number of ways to split should be > 0, but got ", num_split)); 71 ctx, input_shape.dim_size(split_dim) % num_split == 0, 76 "and num_split ", num_split)); 80 const int32 slice_size = input_shape.dim_size(split_dim) / num_split; 97 for (int i = 0; i < num_split; ++i) { 113 const int32 num_split = num_outputs(); variable 136 ctx, num_split > [all...] |
/external/tensorflow/tensorflow/python/kernel_tests/ |
H A D | sparse_split_op_test.py | 81 sp_input=self._SparseTensor_4x6(), num_split=2, axis=0) 98 sp_input=self._SparseTensor_5x7(), num_split=3, axis=1) 116 sp_input=self._SparseTensor_5x7(), num_split=4, axis=1) 137 sp_input=self._SparseTensor_5x7(), num_split=2, axis=0) 152 sp_input=self._SparseTensor_5x7(), num_split=3, axis=0) 173 sp_input=self._SparseTensor_4x6(), num_split=4, axis=0) 195 sp_input=self._SparseTensor_4x6(), num_split=3, axis=1) 213 sp_input=self._SparseTensor_4x6(), num_split=6, axis=1) 242 sp_input=sp_input, num_split=2, axis=1) 253 with self.assertRaisesRegexp(ValueError, 'num_split i [all...] |
H A D | split_op_test.py | 130 num_split = np.random.randint(16, 25) 132 num_split = np.random.randint(2, 8) 133 size_splits = np.random.randint(2, 8, num_split, dtype=np.int32) 140 for i in range(num_split): 157 num_split = 1000 158 size_splits = np.random.randint(1, 3, num_split, dtype=np.int32) 166 for i in range(num_split): 267 num_split = np.random.randint(9, 15) 269 num_split = np.random.randint(2, 8) 270 shape[split_dim] = np.random.randint(2, 5) * num_split [all...] |
/external/tensorflow/tensorflow/core/util/sparse/ |
H A D | sparse_tensor.h | 164 // Split() will split the input SparseTensor into a list of num_split 171 const int num_split); 497 const int num_split) { 501 output_indices.reserve(num_split); 502 output_values.reserve(num_split); 503 output_shapes.reserve(num_split); 507 output_indices_t.reserve(num_split); 508 output_values_t.reserve(num_split); 512 std::vector<int> num_values(num_split, 0); 515 const int split_size = split_dim_size / num_split; 495 Split(const SparseTensor& input_tensor, const int split_dim, const int num_split) argument [all...] |
/external/tensorflow/tensorflow/java/src/test/java/org/tensorflow/ |
H A D | OperationTest.java | 169 private static int split(int[] values, int num_split) { argument 174 .setAttr("num_split", num_split) 180 private static int splitWithInputList(int[] values, int num_split, String name) { argument 185 .setAttr("num_split", num_split)
|
/external/mesa3d/src/gallium/auxiliary/gallivm/ |
H A D | lp_bld_pack.c | 614 int num_split = src_type.width * src_type.length / 128; local 626 assert(num_split <= LP_MAX_VECTOR_WIDTH / 128); 628 for (i = 0; i < num_split / 2; i++) { 639 for (i = 0; i < num_split / 2; i++) { 644 tmpres[i+num_split/2] = lp_build_intrinsic_binary(builder, intrinsic, 648 tmpres[i+num_split/2] = LLVMBuildBitCast(builder, tmpres[i+num_split/2], 652 res = lp_build_concat(gallivm, tmpres, ndst_type, num_split);
|
/external/tensorflow/tensorflow/contrib/lite/toco/tflite/ |
H A D | operator_test.cc | 258 op.num_split = 123; 261 EXPECT_EQ(op.num_split, output_toco_op->num_split);
|
H A D | operator.cc | 613 return ::tflite::CreateSplitOptions(*builder, op.num_split); 618 op->num_split = options.num_splits();
|
/external/tensorflow/tensorflow/python/ops/ |
H A D | sparse_ops.py | 621 num_split=None, 625 """Split a `SparseTensor` into `num_split` tensors along `axis`. 627 If the `sp_input.dense_shape[axis]` is not an integer multiple of `num_split` 628 each slice starting from 0:`shape[axis] % num_split` gets extra one 629 dimension. For example, if `axis = 1` and `num_split = 2` and the 649 num_split: A Python integer. The number of ways to split. 655 `num_split` `SparseTensor` objects resulting from splitting `value`. 665 if num_split is None: 666 raise ValueError("num_split is required") 679 num_split, [all...] |
H A D | array_ops.py | 1319 If `num_or_size_splits` is an integer type, `num_split`, then splits `value` 1320 along dimension `axis` into `num_split` smaller tensors. 1321 Requires that `num_split` evenly divides `value.shape[axis]`. 1367 axis=axis, num_split=num_or_size_splits, value=value, name=name) 1378 num_split=num,
|
/external/tensorflow/tensorflow/contrib/lite/toco/graph_transformations/ |
H A D | propagate_fixed_sizes.cc | 676 CHECK_EQ(split_dim % op->num_split, 0); 677 const int split_depth = split_dim / op->num_split; 682 CHECK_EQ(op->outputs.size(), op->num_split);
|
/external/tensorflow/tensorflow/contrib/estimator/python/estimator/ |
H A D | replicate_model_fn.py | 475 sp_input=tensor, num_split=number_of_shards, axis=0)):
|
/external/tensorflow/tensorflow/contrib/lite/toco/ |
H A D | import_tensorflow.cc | 809 const int num_split = GetIntAttr(node, "num_split"); local 811 for (int i = 1; i < num_split; i++) { 814 op->num_split = num_split;
|
H A D | model.h | 1075 int num_split = 0; member in struct:toco::TensorFlowSplitOperator
|
H A D | export_tensorflow.cc | 1000 (*split_op->mutable_attr())["num_split"].set_i(src_op.num_split); 1302 (*split_op->mutable_attr())["num_split"].set_i(4); // Split into four outputs
|
/external/tensorflow/tensorflow/python/eager/ |
H A D | ops_test.py | 134 split_dim, indices, values, shape, num_split=2)
|
/external/tensorflow/tensorflow/core/graph/ |
H A D | mkl_layout_pass.cc | 1607 int num_split; local 1611 TF_CHECK_OK(GetNodeAttr(orig_node->def(), "num_split", &num_split)); 1616 nb->Attr("num_split", num_split); 3652 int num_split; 3656 TF_CHECK_OK(GetNodeAttr(orig_node->def(), "num_split", &num_split)); 3661 nb->Attr("num_split", num_split); [all...] |
/external/tensorflow/tensorflow/core/ops/ |
H A D | array_ops.cc | 405 .Output("output: num_split * T") 406 .Attr("num_split: int >= 1") 413 int num_split = c->num_outputs(); 426 c->Divide(c->Dim(input, split_dim), num_split, 432 for (int i = 0; i < num_split; ++i) c->set_output(i, out); 440 .Output("output: num_split * T") 441 .Attr("num_split: int >= 1")
|
/external/tensorflow/tensorflow/python/grappler/ |
H A D | layout_optimizer_test.py | 325 value=conv, size_splits=sizes, axis=dim, num_split=3)
|