/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
H A D | pack_op.cc | 47 std::vector<TensorShape> shapes; variable 48 OP_REQUIRES_OK(ctx, ctx->InputList("values", &values, &shapes)); 54 // Verify that all input shapes match 56 OP_REQUIRES(ctx, shapes[0].IsSameSize(shapes[i]), 59 shapes[0].DebugString(), " != values[", i, "].shape = ", 60 shapes[i].DebugString())); 63 int expanded_num_dims = shapes[0].dims() + 1; 74 TensorShape child_shape(shapes[0]);
|
H A D | bcast_ops.cc | 30 // Given shapes of two tensors, computes the broadcast shape. 41 gtl::InlinedVector<BCast::Vec, 2> shapes; variable 49 shapes.push_back(BCast::Vec(shape.begin(), shape.end())); 51 BCast bcast(shapes[0], shapes[1]); 54 "Incompatible shapes: [", str_util::Join(shapes[0], ","), 55 "] vs. [", str_util::Join(shapes[1], ","), "]")); 73 // Given shapes of two tensors, computes the reduction indices for the 90 gtl::InlinedVector<BCast::Vec, 4> shapes; variable [all...] |
/external/tensorflow/tensorflow/contrib/tpu/ops/ |
H A D | outfeed_ops.cc | 68 .Attr("shapes: list(shape)") 72 std::vector<PartialTensorShape> shapes; 74 TF_RETURN_IF_ERROR(c->GetAttr("shapes", &shapes)); 76 if (shapes.size() != dtypes.size()) { 78 "Incorrect number of output shapes specified"); 80 for (int i = 0; i < shapes.size(); ++i) { 82 TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shapes[i], &out)); 94 shapes: The shapes o [all...] |
H A D | infeed_ops.cc | 60 .Attr("shapes: list(shape)") 69 shapes: The shapes of each tensor in `inputs`. 78 .Attr("shapes: list(shape)") 81 std::vector<PartialTensorShape> shapes; 82 TF_RETURN_IF_ERROR(c->GetAttr("shapes", &shapes)); 83 for (int i = 0; i < shapes.size(); ++i) { 85 TF_RETURN_IF_ERROR(c->MakeShapeFromPartialTensorShape(shapes[i], &out)); 96 shapes [all...] |
/external/gemmlowp/meta/generators/ |
H A D | quantized_mul_kernels_arm_32.py | 32 shapes = [(1, 1), (1, 2), (1, 3), (1, 4), (1, 5), (1, 6), (1, 7), (1, 8), 37 shapes)
|
H A D | quantized_mul_kernels_arm_64.py | 32 shapes = [(1, 1), (1, 2), (1, 3), (1, 4), (1, 5), (1, 6), (1, 7), (1, 8), 37 shapes)
|
/external/tensorflow/tensorflow/core/kernels/ |
H A D | sparse_concat_op.cc | 64 OpInputList shapes; variable 65 OP_REQUIRES_OK(context, context->input_list("shapes", &shapes)); 66 OP_REQUIRES(context, shapes.size() == N, 67 errors::InvalidArgument("Expected ", N, " input shapes, got ", 68 shapes.size())); 70 OP_REQUIRES(context, TensorShapeUtils::IsVector(shapes[i].shape()), 72 "Input shapes should be a vector but received shape ", 73 shapes[i].shape().DebugString(), " at position ", i)); 76 const TensorShape input_shape(shapes[ [all...] |
H A D | bcast_ops.cc | 24 // Given shapes of two tensors, computes the broadcast shape. 34 gtl::InlinedVector<BCast::Vec, 4> shapes; variable 44 shapes.push_back(vec); 46 BCast bcast(shapes[0], shapes[1]); 49 "Incompatible shapes: [", str_util::Join(shapes[0], ","), 50 "] vs. [", str_util::Join(shapes[1], ","), "]")); 69 // Given shapes of two tensors, computes the reduction indices for the 83 gtl::InlinedVector<BCast::Vec, 4> shapes; variable [all...] |
/external/tensorflow/tensorflow/python/ops/ |
H A D | data_flow_ops.py | 57 def _as_shape_list(shapes, 61 """Convert shapes to a list of tuples of int (or None).""" 64 if (not isinstance(shapes, collections.Sequence) or not shapes or 65 any(shape is None or isinstance(shape, int) for shape in shapes)): 67 "When providing partial shapes, a list of shapes must be provided.") 68 if shapes is None: 70 if isinstance(shapes, tensor_shape.TensorShape): 71 shapes 244 def shapes(self): member in class:QueueBase 1528 def shapes(self): member in class:BaseStagingArea [all...] |
/external/tensorflow/tensorflow/core/ops/ |
H A D | parsing_ops_test.cc | 65 std::vector<PartialTensorShape> shapes(size); 67 // Make shapes be the sequence [?,1]; [?,1,2], [?,1,2,3]... 70 shapes[i].Clear(); 72 shapes[i].AddDim(-1); 75 shapes[i] = shapes[i - 1]; 77 shapes[i].AddDim(i + 1); 79 if (add_extra_shape) shapes.push_back(PartialTensorShape({})); 80 return shapes; 142 INFER_ERROR("shapes[ [all...] |
/external/tensorflow/tensorflow/python/keras/_impl/keras/utils/ |
H A D | np_utils_test.py | 31 shapes = [(1,), (3,), (4, 3), (5, 4, 3), (3, 1), (3, 2, 1)] 37 labels = [np.random.randint(0, num_classes, shape) for shape in shapes]
|
/external/tensorflow/tensorflow/python/training/ |
H A D | input.py | 181 shapes=[element_shape], 344 # everything in TensorList matches. Maybe just check the inferred shapes? 648 # We want the shapes without the leading batch dimension. 656 def _shapes(tensor_list_list, shapes, enqueue_many): 657 """Calculate and merge the shapes of incoming tensors. 661 shapes: List of shape tuples corresponding to tensors within the lists. 662 enqueue_many: Boolean describing whether shapes will be enqueued as 666 A list of shapes aggregating shape inference info from `tensor_list_list`, 667 or returning `shapes` if it is not `None`. 670 ValueError: If any of the inferred shapes i [all...] |
/external/harfbuzz_ng/src/ |
H A D | gen-arabic-table.py | 152 shapes = {} 185 if items[0] not in shapes: 186 shapes[items[0]] = {} 187 shapes[items[0]][shape] = c 193 keys = shapes.keys () 196 s = [shapes[u][shape] if u in shapes and shape in shapes[u] else 0 212 liga = (shapes[pair[0]]['initial'], shapes[pai [all...] |
/external/tensorflow/tensorflow/contrib/slim/python/slim/data/ |
H A D | prefetch_queue.py | 59 dynamic_pad: Boolean. Whether to allow variable dimensions in input shapes. 79 shapes = [t.get_shape() for t in tensor_list] 83 shapes=shapes,
|
/external/tensorflow/tensorflow/contrib/slim/python/slim/ |
H A D | model_analyzer.py | 65 print('Operations: name -> (type shapes) [size]') 70 shapes = [] 75 shapes.append(tensor_description(output)) 78 print(op.name, '\t->', ', '.join(shapes), '[' + str(op_size) + ']') 84 """Prints the names and shapes of the variables.
|
/external/tensorflow/tensorflow/python/data/util/ |
H A D | sparse.py | 40 def as_dense_shapes(shapes, classes): 41 """Converts sparse tensor shapes to their physical shapes. 44 shapes: a structure of shapes to convert. 48 a structure matching the nested structure of `shapes`, containing 50 `tf.SparseTensor` and matching contents of `shapes` otherwise 52 ret = nest.pack_sequence_as(shapes, [ 54 for shape, c in zip(nest.flatten(shapes), nest.flatten(classes)) 78 def deserialize_sparse_tensors(tensors, types, shapes, classe [all...] |
/external/tensorflow/tensorflow/contrib/training/python/training/ |
H A D | bucket_ops.py | 70 shapes=None, 98 (i) the `shapes` argument is passed, or (ii) all of the tensors in 99 `tensors` must have fully-defined shapes. `ValueError` will be 113 In addition, all output tensors' static shapes, as accessed via the 133 shapes: (Optional) The shapes for each example. Defaults to the 134 inferred shapes for `tensors`. 135 dynamic_pad: Boolean. Allow variable dimensions in input shapes. 137 batch have the same shapes. 155 ValueError: If the `shapes` ar [all...] |
/external/ImageMagick/Magick++/demo/ |
H A D | demos.tap | 20 for demo in button demo flip gravity piddle shapes
|
/external/ImageMagick/PerlMagick/demo/ |
H A D | Makefile | 4 perl shapes.pl
|
/external/tensorflow/tensorflow/contrib/tpu/python/ops/ |
H A D | tpu_ops.py | 77 def infeed_dequeue_tuple(dtypes, shapes, name=None): 83 shapes: A list of shapes (each a `tf.TensorShape` or list of `ints`). 84 The shapes of each tensor in `outputs`. 99 return gen_tpu_ops.infeed_dequeue_tuple(dtypes, shapes, name=name)
|
/external/gemmlowp/test/ |
H A D | benchmark_all_sizes.cc | 278 std::vector<Shape> shapes; local 286 shapes.push_back(shape); 288 // Benchmark an assortment of cubic shapes 294 shapes.push_back(shape); 297 // Benchmark all sorts of shapes 305 shapes.push_back(shape); 310 #error What shapes should we benchmark? (Suggestion: #define BENCHMARK_QUICK) 312 std::shuffle(std::begin(shapes), std::end(shapes), RandomEngine()); 313 return shapes; 317 std::vector<Shape> shapes; local [all...] |
/external/tensorflow/tensorflow/contrib/hvx/hvx_ops_support_checker/ |
H A D | hvx_ops_support_checker_main.cc | 65 std::vector<TensorShape> shapes; local 67 node_def, &data_types, &shapes); 68 if (data_types.empty() || shapes.empty()) { 71 CHECK_EQ(data_types.size(), shapes.size()); 74 << ", " << shapes.at(i).DebugString();
|
/external/tensorflow/tensorflow/contrib/tpu/python/tpu/ |
H A D | tpu_sharding.py | 156 shapes; or shard_index is not None and 219 def get_unsharded_shape(self, shapes): 222 When given a list of shapes of shards, returns the shape of the 227 shapes: The shapes of the Tensor shards to be combined. 233 ValueError: if shapes is not a list of length 234 self.number_of_shards; or any element of shapes is not a valid 236 shapes is not a valid sharding of a full shape. 237 TypeError: if an element of shapes is not convertible to a 241 if len(shapes) ! [all...] |
/external/tensorflow/tensorflow/python/kernel_tests/ |
H A D | cholesky_op_test.py | 197 shapes = self.getShapes([1, 2, 10]) 199 shapes, dtypes=(dtypes_lib.float32, dtypes_lib.float64)) 203 shapes = self.getShapes([1, 2, 10]) 205 shapes, dtypes=(dtypes_lib.complex64, dtypes_lib.complex128)) 209 shapes = self.getShapes([self._backprop_block_size + 1]) 211 shapes, 217 shapes = self.getShapes([2 * self._backprop_block_size + 1]) 219 shapes, dtypes=(dtypes_lib.float32,), scalarTest=True) 223 shapes = self.getShapes([2 * self._backprop_block_size + 1]) 225 shapes, dtype 303 shapes = [ variable in class:CholeskyBenchmark [all...] |
/external/tensorflow/tensorflow/contrib/layers/python/ops/ |
H A D | sparse_feature_cross_op.py | 83 shapes = [sp_input.dense_shape for sp_input in sparse_inputs] 101 shapes, 114 shapes,
|