/external/tensorflow/tensorflow/contrib/reduce_slice_ops/kernels/ |
H A D | reduce_slice_ops.cc | 152 #define REGISTER_CPU_SUMPROD_REDUCE_SLICE_KERNELS(type, index_type) \ 156 .TypeConstraint<index_type>("Tindices"), \ 157 ReduceSliceKernel<CPUDevice, type, index_type, \ 162 .TypeConstraint<index_type>("Tindices"), \ 163 ReduceSliceKernel<CPUDevice, type, index_type, \ 166 #define REGISTER_CPU_MINMAX_REDUCE_SLICE_KERNELS(type, index_type) \ 170 .TypeConstraint<index_type>("Tindices"), \ 171 ReduceSliceKernel<CPUDevice, type, index_type, \ 176 .TypeConstraint<index_type>("Tindices"), \ 177 ReduceSliceKernel<CPUDevice, type, index_type, \ [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | gather_op_helpers.h | 32 // index_type must be must be DT_INT32 or DT_INT64. 40 DataType dtype, DataType index_type,
|
H A D | matrix_band_part_op.cc | 52 DataType index_type = context->input_type(1); variable 62 OP_REQUIRES_OK(context, XlaHelpers::Iota(builder, index_type, m, &iota_m)); 65 OP_REQUIRES_OK(context, XlaHelpers::Iota(builder, index_type, n, &iota_n)); 72 auto zero_index = XlaHelpers::Zero(builder, index_type); 75 XlaHelpers::IntegerLiteral(builder, index_type, m), num_lower); 78 XlaHelpers::IntegerLiteral(builder, index_type, n), num_upper);
|
H A D | index_ops.cc | 62 DataType index_type = output_type(0); local 71 index_type, axis, &output)); 75 index_type, axis, &output));
|
H A D | gather_op.cc | 33 DataType dtype, DataType index_type, 118 auto zero_index = XlaHelpers::Zero(bodyb, index_type); 157 TF_RETURN_IF_ERROR(DataTypeToPrimitiveType(index_type, &ptype)); 194 DataType index_type = input_type(1); variable 195 OP_REQUIRES(context, index_type == DT_INT32 || index_type == DT_INT64, 201 /*indices_are_nd=*/false, input_type(0), index_type, 29 XlaGather(const xla::ComputationDataHandle& input, const TensorShape& input_shape, const xla::ComputationDataHandle& indices, TensorShape indices_shape, int64 axis, bool indices_are_nd, DataType dtype, DataType index_type, xla::ComputationBuilder* builder, xla::ComputationDataHandle* gather_output) argument
|
H A D | variable_ops.cc | 119 DataType index_type = ctx->input_type(1); variable 123 /*axis=*/0, /*indices_are_nd=*/false, type, index_type,
|
/external/tensorflow/tensorflow/core/kernels/ |
H A D | one_hot_op.cc | 128 #define REGISTER_ONE_HOT_INDEX(type, index_type) \ 131 .TypeConstraint<index_type>("TI") \ 134 OneHotOp<CPUDevice, type, index_type>); 171 #define REGISTER_ONE_HOT_GPU_INDEX(type, index_type) \ 174 .TypeConstraint<index_type>("TI") \ 177 OneHotOp<GPUDevice, type, index_type>);
|
H A D | gather_op.cc | 122 #define REGISTER_GATHER_FULL(dev, type, index_type) \ 126 .TypeConstraint<index_type>("Tindices"), \ 127 GatherOp<dev##Device, type, index_type>); \ 131 .TypeConstraint<index_type>("Tindices") \ 133 GatherOp<dev##Device, type, index_type>)
|
H A D | segment_reduction_ops.cc | 298 #define REGISTER_CPU_KERNEL_SEGMENT(name, functor, type, index_type, \ 304 .TypeConstraint<index_type>("Tindices"), \ 305 SegmentReductionOp<CPUDevice, type, index_type, functor, default_value>) 307 #define REGISTER_REAL_CPU_KERNELS(type, index_type) \ 309 type, index_type, 0); \ 311 "SegmentMean", Eigen::internal::MeanReducer<type>, type, index_type, 0); \ 313 "SegmentProd", Eigen::internal::ProdReducer<type>, type, index_type, 1); \ 315 type, index_type, 0); \ 317 type, index_type, 0) 319 #define REGISTER_COMPLEX_CPU_KERNELS(type, index_type) \ [all...] |
H A D | sparse_to_dense_op.cc | 140 #define REGISTER_KERNELS(type, index_type) \ 144 .TypeConstraint<index_type>("Tindices"), \ 145 SparseToDense<type, index_type>);
|
H A D | gather_nd_op_test.cc | 60 void MakeOp(DataType index_type) { argument 63 .Input(FakeInput(index_type))
|
H A D | sequence_ops_test.cc | 45 void MakeOp(DataType input_type, DataType index_type) { argument 49 .Input(FakeInput(index_type))
|
H A D | scatter_nd_op_test.cc | 40 void MakeOp(DataType variable_ref_type, DataType index_type) { argument 43 .Input(FakeInput(index_type)) 246 void MakeBenchmarkOp(const char* op, DataType index_type) { argument 249 .Input(FakeInput(index_type))
|
H A D | scatter_nd_op.cc | 174 #define REGISTER_SCATTER_ND_KERNEL_INDEX(type, index_type, dev, name) \ 178 .TypeConstraint<index_type>("Tindices") \ 180 ScatterNdOp<dev##Device, type, index_type>) 182 #define REGISTER_SCATTER_ND_UPDATE_KERNEL_INDEX(type, index_type, dev, name, \ 188 .TypeConstraint<index_type>("Tindices"), \ 189 ScatterNdUpdateOp<dev##Device, type, index_type, op>) 191 #define REGISTER_RESOURCE_SCATTER_ND_UPDATE_KERNEL_INDEX(type, index_type, \ 197 .TypeConstraint<index_type>("Tindices") \ 199 ScatterNdUpdateOp<dev##Device, type, index_type, op>)
|
H A D | scatter_op_test.cc | 40 void MakeOp(DataType variable_ref_type, DataType index_type) { argument 43 .Input(FakeInput(index_type)) 229 void MakeBenchmarkOp(const char* op, DataType index_type) { argument 232 .Input(FakeInput(index_type))
|
H A D | gather_nd_op.cc | 55 #define REGISTER_GATHER_ND_FULL(dev, type, index_type) \ 59 .TypeConstraint<index_type>("Tindices"), \ 60 GatherNdOp<dev##Device, type, index_type>)
|
H A D | gather_op_test.cc | 42 void MakeOp(DataType data_type, DataType index_type) { argument 45 .Input(FakeInput(index_type)) 46 .Input(FakeInput(index_type))
|
H A D | sparse_to_dense_op_test.cc | 41 void MakeOp(int dim, DataType index_type, DataType value_type) { argument 43 .Input(FakeInput(index_type)) 44 .Input(FakeInput(index_type))
|
H A D | resource_variable_ops.cc | 535 #define REGISTER_GATHER_FULL(dev, type, index_type) \ 540 .TypeConstraint<index_type>("Tindices"), \ 541 ResourceGatherOp<dev##Device, type, index_type>) 614 #define REGISTER_SCATTER_KERNEL_INDEX(type, index_type, dev, name, op) \ 620 .TypeConstraint<index_type>("Tindices"), \ 621 ResourceScatterUpdateOp<dev##Device, type, index_type, op>)
|
/external/skia/src/sksl/ir/ |
H A D | SkSLIndexExpression.h | 20 static const Type& index_type(const Context& context, const Type& type) { function in namespace:SkSL 55 : INHERITED(base->fOffset, kIndex_Kind, index_type(context, base->fType))
|
/external/skqp/src/sksl/ir/ |
H A D | SkSLIndexExpression.h | 20 static const Type& index_type(const Context& context, const Type& type) { function in namespace:SkSL 55 : INHERITED(base->fOffset, kIndex_Kind, index_type(context, base->fType))
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
H A D | gather.cc | 88 #define TF_LITE_GATHER(data_type, index_type) \ 91 GetTensorData<index_type>(positions), GetTensorDims(positions), \
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
H A D | xla_helpers.h | 105 DataType index_type, const TensorShape& indices_shape,
|
H A D | xla_helpers.cc | 227 int axis, DataType index_type, 244 switch (index_type) { 256 DataTypeString(index_type)); 226 OneHot(xla::ComputationBuilder* builder, int64 depth, int axis, DataType index_type, const TensorShape& indices_shape, const xla::ComputationDataHandle& indices, const xla::ComputationDataHandle& on_value, const xla::ComputationDataHandle& off_value, xla::ComputationDataHandle* one_hot) argument
|
/external/tensorflow/tensorflow/python/training/ |
H A D | training_ops_test.py | 208 for (dtype, index_type) in itertools.product( 217 indices = np.array([0, 2]).astype(index_type) 221 for (dtype, index_type) in itertools.product( 230 indices = np.array([0, 2]).astype(index_type) 234 for (dtype, index_type) in itertools.product( 245 indices = np.array([0, 2]).astype(index_type)
|