Searched refs:reduced (Results 1 - 25 of 45) sorted by relevance

12

/external/clang/test/CodeGen/
H A Dregparm.c14 extern void FASTCALL reduced(char b, double c, foo* d, double e, int f);
23 // CHECK: call void @reduced(i8 inreg signext 0, {{.*}} %struct.foo* inreg null
24 reduced(0, 0.0, 0, 0.0, 0);
/external/python/cpython2/Lib/lib2to3/
H A Dbtm_utils.py130 reduced = reduce_tree(child, new_node)
131 if reduced is not None:
132 new_node.children.append(reduced)
138 reduced = reduce_tree(child, new_node)
139 if reduced:
140 new_node.children.append(reduced)
142 # delete the group if all of the children were reduced to None
230 reduced = reduce_tree(child, new_node)
231 if reduced is not None:
232 new_node.children.append(reduced)
[all...]
/external/python/cpython3/Lib/lib2to3/
H A Dbtm_utils.py129 reduced = reduce_tree(child, new_node)
130 if reduced is not None:
131 new_node.children.append(reduced)
137 reduced = reduce_tree(child, new_node)
138 if reduced:
139 new_node.children.append(reduced)
141 # delete the group if all of the children were reduced to None
229 reduced = reduce_tree(child, new_node)
230 if reduced is not None:
231 new_node.children.append(reduced)
[all...]
/external/tensorflow/tensorflow/compiler/xla/service/
H A Dreduce_precision_insertion.cc96 HloInstruction* reduced = instruction->parent()->AddInstruction( local
99 TF_RETURN_IF_ERROR(instruction->ReplaceAllUsesWith(reduced));
137 HloInstruction* reduced = nullptr; local
143 reduced = user;
148 if (!reduced) {
149 reduced = instruction->parent()->AddInstruction(
154 TF_RETURN_IF_ERROR(instruction->ReplaceOperandWith(i, reduced));
/external/clang/utils/analyzer/
H A Dreducer.pl64 `cp $srcFile $file.reduced`;
65 print "$prog: generated '$file.reduced";
/external/tensorflow/tensorflow/python/kernel_tests/
H A Dreduce_join_op_test.py113 keep_dims: Whether or not to retain reduced dimensions.
263 reduced = string_ops.reduce_join(placeholder, reduction_indices=0)
264 output_array = reduced.eval(feed_dict={placeholder.name: input_array})
266 self.assertAllEqual(truth_shape, reduced.get_shape())
276 reduced = string_ops.reduce_join(
278 output_array_dim_zero = reduced.eval(feed_dict={placeholder.name: [0]})
279 output_array_dim_one = reduced.eval(feed_dict={placeholder.name: [1]})
282 self.assertAllEqual(truth_shape, reduced.get_shape())
368 reduced = string_ops.reduce_join(
372 reduced
[all...]
H A Dconv1d_test.py47 reduced = array_ops.squeeze(c)
48 output = reduced.eval()
H A Dlinalg_grad_test.py50 reduced = math_ops.reduce_sum(determinants)
51 sum_grad = gradients_impl.gradients(reduced, batch_identity)[0]
H A Dsparse_ops_test.py694 reduced = sparse_ops.sparse_reduce_sum(sp_t, axes)
697 reduced,
698 reduced.eval().shape)
702 reduced = sparse_ops.sparse_reduce_sum(sp_t, -1)
704 reduced,
705 reduced.eval().shape)
/external/skia/tests/
H A DClipStackTest.cpp860 // reduced stack.
1031 // Get the reduced version of the stack.
1034 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, queryBounds, caps); local
1037 reduced->maskElements().isEmpty() ||
1038 SkClipStack::kInvalidGenID != reduced->maskGenID(),
1041 if (!reduced->maskElements().isEmpty()) {
1042 REPORTER_ASSERT(reporter, reduced->hasScissor(), testCase.c_str());
1046 REPORTER_ASSERT(reporter, reduced->maskRequiresAA() == doAA, testCase.c_str());
1049 // Build a new clip stack based on the reduced clip elements
1051 if (GrReducedClip::InitialState::kAllOut == reduced
1099 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, bounds, caps); local
[all...]
H A DPathOpsConicLineIntersectionTest.cpp110 SkPoint reduced[3]; local
113 SkPath::Verb order1 = SkReduceOrder::Conic(floatConic, reduced);
/external/skqp/tests/
H A DClipStackTest.cpp860 // reduced stack.
1031 // Get the reduced version of the stack.
1034 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, queryBounds, caps); local
1037 reduced->maskElements().isEmpty() ||
1038 SkClipStack::kInvalidGenID != reduced->maskGenID(),
1041 if (!reduced->maskElements().isEmpty()) {
1042 REPORTER_ASSERT(reporter, reduced->hasScissor(), testCase.c_str());
1046 REPORTER_ASSERT(reporter, reduced->maskRequiresAA() == doAA, testCase.c_str());
1049 // Build a new clip stack based on the reduced clip elements
1051 if (GrReducedClip::InitialState::kAllOut == reduced
1099 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, bounds, caps); local
[all...]
H A DPathOpsConicLineIntersectionTest.cpp110 SkPoint reduced[3]; local
113 SkPath::Verb order1 = SkReduceOrder::Conic(floatConic, reduced);
/external/eigen/unsupported/test/
H A Dcxx11_tensor_argmax.cpp64 Tensor<Tuple<DenseIndex, float>, 0, DataLayout> reduced; local
66 reduced = index_tuples.reduce(
71 VERIFY_IS_EQUAL(maxi(), reduced(0).second);
96 Tensor<Tuple<DenseIndex, float>, 0, DataLayout> reduced; local
98 reduced = index_tuples.reduce(
103 VERIFY_IS_EQUAL(mini(), reduced(0).second);
190 // Expect max to be in the first index of the reduced dimension
212 // Expect max to be in the last index of the reduced dimension
248 // Expect min to be in the first index of the reduced dimension
270 // Expect min to be in the last index of the reduced dimensio
[all...]
/external/gemmlowp/meta/generators/
H A Dstreams_common.py126 reduced = aggregators[:reduced_count]
128 emitter.EmitVSumReduce('u32', len(aggregators), 4, reduced, aggregators)
130 for temp in reduced:
133 for temp in reduced:
136 emitter.EmitVStoreA(1, 32, reduced,
H A Dquantized_mul_kernels_common.py521 reduced = aggregators[:reduced_count]
522 emitter.EmitVSumReduce('u32', len(aggregators), 4, reduced, aggregators)
523 return reduced
572 reduced = _GenerateAggregatorReduce(self.asm_emitter, aggregators, kernel_m,
576 registers, reduced,
/external/turbine/java/com/google/turbine/deps/
H A DDependencies.java142 Set<String> reduced = new HashSet<>(directJarsToTargets.keySet());
154 reduced.add(dep.getPath());
165 return Collections2.filter(transitiveClasspath, Predicates.in(reduced));
/external/tensorflow/tensorflow/core/kernels/
H A Dreduction_ops_common.h125 // Permutation of reduced dims needed to put reduction dimensions at the end
175 // A temporary tensor whose size matches the size of the reduced
218 // all reduced dimensions are last and reuse the 2-D -> 1-D case.
228 const int64 reduced = shuffled.NumElements() / unreduced; variable
231 const_shuffled.shaped<T, 2>({unreduced, reduced}),
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/streams/
H A Drewrite.rb213 reduced = []
257 reduced.unshift( operation ) if operation
260 @operations.replace( reduced )
/external/capstone/
H A Dconfig.mk25 # such as @regs_read/write & @group. The amount of binary size reduced is
/external/eigen/Eigen/src/Core/arch/CUDA/
H A DPacketMathHalf.h865 float reduced = predux<Packet8f>(af);
866 return Eigen::half(reduced);
871 float reduced = predux_max<Packet8f>(af);
872 return Eigen::half(reduced);
877 float reduced = predux_min<Packet8f>(af);
878 return Eigen::half(reduced);
883 float reduced = predux_mul<Packet8f>(af);
884 return Eigen::half(reduced);
/external/libjpeg-turbo/simd/
H A Djidctred-mmx.asm2 ; jidctred.asm - reduced-size IDCT (MMX)
16 ; This file contains inverse-DCT routines that produce reduced-size
98 ; producing a reduced-size 4x4 output block.
492 ; producing a reduced-size 2x2 output block.
H A Djidctred-sse2-64.asm2 ; jidctred.asm - reduced-size IDCT (64-bit SSE2)
17 ; This file contains inverse-DCT routines that produce reduced-size
99 ; producing a reduced-size 4x4 output block.
402 ; producing a reduced-size 2x2 output block.
H A Djidctred-sse2.asm2 ; jidctred.asm - reduced-size IDCT (SSE2)
16 ; This file contains inverse-DCT routines that produce reduced-size
98 ; producing a reduced-size 4x4 output block.
413 ; producing a reduced-size 2x2 output block.
/external/tensorflow/tensorflow/cc/gradients/
H A Dmath_grad.cc607 // input_shape: 1-D Tensor, the shape of the Tensor being reduced.
664 // The partial derivative for any input along a "reduced" dimension
714 // all gradients are also divided by the size of reduced groups.
769 // The partial derivative for any input along a "reduced" dimension
901 auto reduced = Cast(cpu_scope, reduction_indices_pos, DataType::DT_INT32); local
907 auto other = SetDiff1D(cpu_scope, idx, reduced).out;
911 Concat(cpu_scope, std::initializer_list<Input>{reduced, other}, 0);
914 auto reduced_num = Prod(cpu_scope, Gather(scope, input_shape, reduced), 0);

Completed in 5919 milliseconds

12