/external/tensorflow/tensorflow/compiler/xla/ |
H A D | layout_util.cc | 42 // minor_to_major to the value that represents the default layout. 45 minor_to_major) { 49 const int64 size = minor_to_major->size(); 51 minor_to_major->Set(i, size - 1 - i); 58 tensorflow::gtl::ArraySlice<int64> minor_to_major) { 61 for (int64 dimension_number : minor_to_major) { 81 minor_to_major = layout.mutable_minor_to_major(); local 82 minor_to_major->Resize(rank, 0); 83 SetDefaultLayoutToContainer(minor_to_major); 123 minor_to_major local 43 SetDefaultLayoutToContainer( tensorflow::protobuf::RepeatedField<tensorflow::protobuf_int64>* minor_to_major) argument 57 MakeLayout( tensorflow::gtl::ArraySlice<int64> minor_to_major) argument [all...] |
H A D | shape_util.cc | 117 // Constructs and returns the new shape with the given minor_to_major order in 121 tensorflow::gtl::ArraySlice<int64> minor_to_major) { 122 if (dimensions.size() != minor_to_major.size()) { 124 dimensions.size(), minor_to_major.size()); 133 for (int64 value : minor_to_major) { 193 tensorflow::gtl::ArraySlice<int64> minor_to_major) { 194 return MakeShapeWithLayoutInternal(element_type, dimensions, minor_to_major) 853 // Construct minor_to_major of stripped shape. The order of the non-degenerate 855 // vector of the non-degenerate dimensions from the original minor_to_major 857 std::vector<int64> minor_to_major; local 119 MakeShapeWithLayoutInternal( PrimitiveType element_type, tensorflow::gtl::ArraySlice<int64> dimensions, tensorflow::gtl::ArraySlice<int64> minor_to_major) argument 191 MakeShapeWithLayout( PrimitiveType element_type, tensorflow::gtl::ArraySlice<int64> dimensions, tensorflow::gtl::ArraySlice<int64> minor_to_major) argument [all...] |
H A D | layout_util.h | 37 static Layout MakeLayout(tensorflow::gtl::ArraySlice<int64> minor_to_major); 135 // Returns the minor_to_major array for the given Shape. Requires that the
|
H A D | shape_util_test.cc | 648 F32, {3, 4}, AsInt64Slice(input_layout.minor_to_major())), 650 F32, {6, 2}, AsInt64Slice(output_layout.minor_to_major()))), 674 EXPECT_THAT(aligned_shape.value().layout().minor_to_major(), 681 EXPECT_THAT(aligned_shape.value().layout().minor_to_major(), 693 EXPECT_THAT(aligned_shape.value().layout().minor_to_major(),
|
H A D | layout_util_test.cc | 31 tensorflow::gtl::ArraySlice<int64> minor_to_major) { 33 *shape.mutable_layout() = LayoutUtil::MakeLayout(minor_to_major); 217 ::testing::ContainsRegex("layout minor_to_major field contains .* " 29 MakeShapeWithLayout(PrimitiveType element_type, tensorflow::gtl::ArraySlice<int64> dimensions, tensorflow::gtl::ArraySlice<int64> minor_to_major) argument
|
H A D | shape_util.h | 323 // Constructs a new shape with the given minor_to_major order in its Layout. 327 tensorflow::gtl::ArraySlice<int64> minor_to_major); 465 // stripped minor_to_major preserves the relative ordering of non-degenerate 470 // input shape: F32 [1, 2, 1], minor_to_major = {0, 1, 2} 471 // stripped shape: F32 [2], minor_to_major = {0} 473 // input shape: F32 [6, 1, 5], minor_to_major = {2, 0, 1} 474 // stripped shape: F32 [6, 5], minor_to_major = {1, 0} 476 // input shape: F32 [1, 7, 1, 6, 5, 1], minor_to_major = {0, 2, 5, 4, 3, 1} 477 // stripped shape: F32 [7, 6, 5], minor_to_major = {0, 2, 1} 479 // input shape: F32 [1, 1], minor_to_major [all...] |
H A D | xla_data.proto | 144 repeated int64 minor_to_major = 1;
|
H A D | literal_util.h | 906 AsInt64Slice(layout.minor_to_major()))); 1008 AsInt64Slice(layout.minor_to_major())));
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | layout_assignment_test.cc | 65 for (auto& minor_to_major : minor_to_majors) { 77 Layout layout = LayoutUtil::MakeLayout(minor_to_major); 134 for (auto& minor_to_major : minor_to_majors) { 137 {{1.0, 2.0}, {3.0, 4.0}}, LayoutUtil::MakeLayout(minor_to_major)); 139 {{5.0, 6.0}, {7.0, 8.0}}, LayoutUtil::MakeLayout(minor_to_major)); 159 Layout layout = LayoutUtil::MakeLayout(minor_to_major); 350 AsInt64Slice(log->shape().layout().minor_to_major()); 355 AsInt64Slice(reshape->shape().layout().minor_to_major()); 429 EXPECT_THAT(broadcast->shape().layout().minor_to_major(), 482 EXPECT_THAT(broadcast->shape().layout().minor_to_major(), ElementsAr [all...] |
/external/tensorflow/tensorflow/compiler/xla/tests/ |
H A D | client_library_test_base.cc | 172 std::vector<int64> minor_to_major(ShapeUtil::Rank(expected.shape())); 173 std::iota(minor_to_major.begin(), minor_to_major.end(), 0); 177 AsInt64Slice(expected.shape().dimensions()), minor_to_major); 183 } while (std::next_permutation(minor_to_major.begin(), minor_to_major.end())); 216 std::vector<int64> minor_to_major(ShapeUtil::Rank(literal->shape())); 217 std::iota(minor_to_major.begin(), minor_to_major.end(), 0); 220 literal->Relayout(LayoutUtil::MakeLayout(minor_to_major)); [all...] |
H A D | matrix_ops_simple_test.cc | 171 auto minor_to_major = [](bool row_major) -> std::vector<int64> { local 184 lhs, LayoutUtil::MakeLayout(minor_to_major(row_major))))); 188 rhs, LayoutUtil::MakeLayout(minor_to_major(row_major))))); 237 auto minor_to_major = [](bool row_major) -> std::vector<int64> { local 251 lhs, LayoutUtil::MakeLayout(minor_to_major(row_major))))); 256 rhs, LayoutUtil::MakeLayout(minor_to_major(row_major)))));
|
H A D | broadcast_simple_test.cc | 59 tensorflow::gtl::ArraySlice<int64> minor_to_major, Shape* r3_shape, 61 *r3_shape = ShapeUtil::MakeShapeWithLayout(F32, bounds, minor_to_major); 64 LayoutUtil::MakeLayout(minor_to_major)); 72 tensorflow::gtl::ArraySlice<int64> minor_to_major, Shape* r2_shape, 74 *r2_shape = ShapeUtil::MakeShapeWithLayout(F32, bounds, minor_to_major); 77 LayoutUtil::MakeLayout(minor_to_major)); 57 MakeR3Data( tensorflow::gtl::ArraySlice<int64> bounds, tensorflow::gtl::ArraySlice<int64> minor_to_major, Shape* r3_shape, Array3D<float>* r3_array, float start, float end, int seed) argument 70 MakeR2Data( tensorflow::gtl::ArraySlice<int64> bounds, tensorflow::gtl::ArraySlice<int64> minor_to_major, Shape* r2_shape, Array2D<float>* r2_array, float start, float end, int seed) argument
|
H A D | params_test.cc | 496 original.layout().minor_to_major().begin(), 497 original.layout().minor_to_major().end());
|
H A D | literal_test_util.cc | 758 tensorflow::gtl::ArraySlice<int64> minor_to_major, const Literal& literal) { 764 CHECK_EQ(new_dimensions.size(), minor_to_major.size()); 773 *shape_with_layout.mutable_layout() = LayoutUtil::MakeLayout(minor_to_major); 756 Reshape( tensorflow::gtl::ArraySlice<int64> new_dimensions, tensorflow::gtl::ArraySlice<int64> minor_to_major, const Literal& literal) argument
|
H A D | literal_test_util.h | 201 // buffer of the input literal is assumed to have the given minor_to_major 205 tensorflow::gtl::ArraySlice<int64> minor_to_major,
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
H A D | shape_partition.cc | 29 const int64 dimension = shape_.layout().minor_to_major(i); 107 dimensions_[i] = shape_.layout().minor_to_major(
|
H A D | shape_partition_test.cc | 213 const int64 dimension = shape.layout().minor_to_major(
|
H A D | ir_emitter.cc | 1310 int64 operand_dim = operand_shape.layout().minor_to_major(operand_dim_idx); 1313 result_shape.layout().minor_to_major(result_dim_idx++)) {
|
/external/tensorflow/tensorflow/compiler/xla/python/ |
H A D | numpy_bridge.cc | 239 TF_ASSIGN_OR_RETURN(py_minor_to_major, call_method("minor_to_major")); 245 "Return value of Shape method minor_to_major() is neither a tuple " 252 "Shape methods dimensions() and minor_to_major() return " 256 std::vector<int64> minor_to_major(length); 264 minor_to_major[i] = 266 if (minor_to_major[i] == -1 && PyErr_Occurred()) { 276 minor_to_major);
|
H A D | xla_client.py | 191 def __init__(self, np_dtype, dimensions, minor_to_major=None): 195 self._minor_to_major = minor_to_major 206 'minor_to_major={!r})').format(self.np_dtype, self._dimensions, 220 def minor_to_major(self): member in class:Shape 256 def update_minor_to_major(self, minor_to_major): 257 if not isinstance(minor_to_major, tuple): 258 raise TypeError('minor_to_major must be a tuple') 259 updated = Shape(self.np_dtype, tuple(self.dimensions()), minor_to_major)
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
H A D | ir_array.cc | 43 int64 dimension = layout_.minor_to_major(i);
|