/external/tensorflow/tensorflow/python/ops/ |
H A D | array_grad.py | 37 def _PackGrad(op, grad): 38 """Gradient for pack op.""" 39 return array_ops.unstack(grad, num=op.get_attr("N"), axis=op.get_attr("axis")) 43 def _UnpackGrad(op, *grads): 44 """Gradient for unpack op.""" 45 return array_ops.stack(grads, axis=op.get_attr("axis")) 48 def _ConcatGradHelper(op, grad, start_value_index, end_value_index, dim_index): 49 """Gradient for concat op. 52 op [all...] |
/external/clang/test/CodeGen/ |
H A D | asm_arm64.c | 28 void t4(long op) { argument 30 asm ("mov x0, %1; svc #0;" : "=r"(x1) :"r"(op),"r"(x1) :"x0" );
|
/external/parameter-framework/asio/include/asio/detail/impl/ |
H A D | task_io_service.hpp | 38 typedef completion_handler<Handler> op; typedef 39 typename op::ptr p = { asio::detail::addressof(handler), 41 sizeof(op), handler), 0 }; 42 p.p = new (p.v) op(handler); 58 typedef completion_handler<Handler> op; typedef 59 typename op::ptr p = { asio::detail::addressof(handler), 61 sizeof(op), handler), 0 }; 62 p.p = new (p.v) op(handler);
|
/external/parameter-framework/asio-1.10.6/include/asio/detail/impl/ |
H A D | task_io_service.hpp | 38 typedef completion_handler<Handler> op; typedef 39 typename op::ptr p = { asio::detail::addressof(handler), 41 sizeof(op), handler), 0 }; 42 p.p = new (p.v) op(handler); 58 typedef completion_handler<Handler> op; typedef 59 typename op::ptr p = { asio::detail::addressof(handler), 61 sizeof(op), handler), 0 }; 62 p.p = new (p.v) op(handler);
|
/external/python/cpython2/Misc/ |
H A D | pymemcompat.h | 81 #define PyObject_GC_Track(op) 82 #define PyObject_GC_UnTrack(op)
|
/external/skia/src/core/ |
H A D | SkRecordDraw.cpp | 89 DRAW(ClipPath, clipPath(r.path, r.opAA.op(), r.opAA.aa())); 90 DRAW(ClipRRect, clipRRect(r.rrect, r.opAA.op(), r.opAA.aa())); 91 DRAW(ClipRect, clipRect(r.rect, r.opAA.op(), r.opAA.aa())); 92 DRAW(ClipRegion, clipRegion(r.region, r.op)); 151 // don't execute this op, pixels in this rectangle might draw incorrectly. So 189 template <typename T> void operator()(const T& op) { argument 190 this->updateCTM(op); 191 this->updateClipBounds(op); 192 this->trackBounds(op); 240 void updateCTM(const Restore& op) { fCT argument 241 updateCTM(const SetMatrix& op) argument 242 updateCTM(const Concat& op) argument 243 updateCTM(const Translate& op) argument 249 updateClipBounds(const ClipPath& op) argument 250 updateClipBounds(const ClipRRect& op) argument 251 updateClipBounds(const ClipRect& op) argument 252 updateClipBounds(const ClipRegion& op) argument 267 updateClipBounds(const Restore& op) argument 282 updateClipBounds(const SaveLayer& op) argument 292 trackBounds(const SaveLayer& op) argument 305 trackBounds(const T& op) argument [all...] |
/external/skqp/src/core/ |
H A D | SkRecordDraw.cpp | 89 DRAW(ClipPath, clipPath(r.path, r.opAA.op(), r.opAA.aa())); 90 DRAW(ClipRRect, clipRRect(r.rrect, r.opAA.op(), r.opAA.aa())); 91 DRAW(ClipRect, clipRect(r.rect, r.opAA.op(), r.opAA.aa())); 92 DRAW(ClipRegion, clipRegion(r.region, r.op)); 151 // don't execute this op, pixels in this rectangle might draw incorrectly. So 189 template <typename T> void operator()(const T& op) { argument 190 this->updateCTM(op); 191 this->updateClipBounds(op); 192 this->trackBounds(op); 240 void updateCTM(const Restore& op) { fCT argument 241 updateCTM(const SetMatrix& op) argument 242 updateCTM(const Concat& op) argument 243 updateCTM(const Translate& op) argument 249 updateClipBounds(const ClipPath& op) argument 250 updateClipBounds(const ClipRRect& op) argument 251 updateClipBounds(const ClipRect& op) argument 252 updateClipBounds(const ClipRegion& op) argument 267 updateClipBounds(const Restore& op) argument 282 updateClipBounds(const SaveLayer& op) argument 292 trackBounds(const SaveLayer& op) argument 305 trackBounds(const T& op) argument [all...] |
/external/syslinux/gpxe/src/drivers/bitbash/ |
H A D | bitbash.c | 41 basher->op->write ( basher, bit_id, ( data ? -1UL : 0 ) ); 56 return ( basher->op->read ( basher, bit_id ) ? -1UL : 0 );
|
/external/tensorflow/tensorflow/contrib/lite/toco/graph_transformations/ |
H A D | remove_unused_op.cc | 30 const auto* op = it->get(); local 35 for (const auto& output : op->outputs) { 50 CHECK(op->type == OperatorType::kFill); 78 if (op->unresolved_outputs) { 80 LogName(*op)); 85 LogName(*op)); 92 for (const auto& input : op->inputs) { 101 for (const auto& output : op->outputs) {
|
H A D | remove_trivial_slice.cc | 31 bool IsSliceTrivial(const Model& model, const Operator& op, argument 33 CHECK(op.type == OperatorType::kSlice); 36 const auto& input_array = model.GetArray(op.inputs[0]); 37 const auto& output_array = model.GetArray(op.outputs[0]); 42 LogName(op));
|
/external/tensorflow/tensorflow/contrib/tensor_forest/kernels/v4/ |
H A D | leaf_model_operators_test.cc | 65 void TestClassificationNormalUse(const std::unique_ptr<LeafModelOperator>& op) { argument 67 op->InitModel(&l); 70 EXPECT_EQ(op->GetOutputValue(l, i), 0); 79 op->UpdateModel(&l, target.get(), 0); 80 op->UpdateModel(&l, target.get(), 1); 81 op->UpdateModel(&l, target.get(), 2); 83 EXPECT_FLOAT_EQ(op->GetOutputValue(l, 1), 3.4); 89 std::unique_ptr<LeafModelOperator> op( 91 TestClassificationNormalUse(op); 97 std::unique_ptr<LeafModelOperator> op( [all...] |
/external/tensorflow/tensorflow/core/api_def/ |
H A D | update_api_def.h | 25 // constructed based on the given op. 26 string CreateApiDef(const OpDef& op); 28 // Removes .Doc call for the given op. 32 string RemoveDoc(const OpDef& op, const string& file_contents,
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
H A D | bucketize_op_test.py | 30 op = math_ops._bucketize( 35 self.assertAllEqual(expected_out, sess.run(op)) 38 op = math_ops._bucketize( 43 self.assertAllEqual(expected_out, sess.run(op)) 46 op = math_ops._bucketize( 51 self.assertAllEqual(expected_out, sess.run(op)) 54 op = math_ops._bucketize( 59 sess.run(op)
|
/external/valgrind/none/tests/mips64/ |
H A D | macro_fpu.h | 56 #define UNOPdd(op) \ 59 op" %1, %2" "\n\t" \ 65 #define UNOPff(op) \ 68 op" %1, %2" "\n\t" \ 74 #define UNOPfd(op) \ 77 op" %1, %2" "\n\t" \ 83 #define UNOPdf(op) \ 86 op" %1, %2" "\n\t" \ 92 #define UNOPfw(op) \ 95 op" [all...] |
/external/python/cpython3/Lib/ |
H A D | opcode.py | 37 opname = ['<%r>' % (op,) for op in range(256)] 39 def def_op(name, op): 40 opname[op] = name 41 opmap[name] = op 43 def name_op(name, op): 44 def_op(name, op) 45 hasname.append(op) 47 def jrel_op(name, op): 48 def_op(name, op) [all...] |
/external/elfutils/libdw/ |
H A D | dwarf_getlocation_die.c | 1 /* Return DIE associated with a location expression op. 37 dwarf_getlocation_die (Dwarf_Attribute *attr, const Dwarf_Op *op, argument 44 switch (op->atom) 48 dieoff = op->number; 57 dieoff = attr->cu->start + op->number; 62 dieoff = attr->cu->start + op->number2;
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | reshape_mover_test.cc | 33 namespace op = xla::testing::opcode_matchers; 56 op::Add(op::Reshape(param0), op::Reshape(param1))); 61 op::Add(op::Reshape(param0), op::Reshape(param1))); 94 op::Add(op::Reshape(rng0), const1)); 99 op [all...] |
/external/tensorflow/tensorflow/contrib/cudnn_rnn/ops/ |
H A D | cudnn_rnn_ops_test.cc | 17 #include "tensorflow/core/framework/op.h" 28 ShapeInferenceTestOp op("CudnnRNNParamsSize"); 29 INFER_OK(op, "[1];[1];[1]", "[1]"); 33 ShapeInferenceTestOp op("CudnnRNN"); 42 .Finalize(&op.node_def)); 60 INFER_OK(op, input_shapes_desc, output_shapes_desc);
|
/external/tensorflow/tensorflow/contrib/kfac/python/ops/ |
H A D | op_queue.py | 15 """Helper for choosing which op to run next in a distributed setting.""" 42 self._ops_by_name = {op.name: op for op in ops} 45 op_names = tf_ops.convert_to_tensor(list(sorted(op.name for op in ops))) 56 """Chooses which op to run next.
|
/external/tensorflow/tensorflow/contrib/tensor_forest/hybrid/python/ops/ |
H A D | training_ops.py | 46 def _RoutingFunctionGradient(op, grad): 50 op: The RoutingFunction op. 51 grad: Gradient with respect to the output of the RoutingFunction op. 54 Gradients with respect to the input of the RoutingFunction op. 58 input_data_tensor = op.inputs[0] 59 tree_weights_tensor = op.inputs[1] 60 tree_thresholds_tensor = op.inputs[2] 62 routing_function_tensor = op.outputs[0] 87 max_nodes=op [all...] |
/external/tensorflow/tensorflow/core/ops/ |
H A D | remote_fused_graph_ops_test.cc | 18 #include "tensorflow/core/framework/op.h" 28 ShapeInferenceTestOp op("RemoteFusedGraphExecute"); 29 auto set_n = [&op](int input1_count, int input2_count, int output_count) { 48 TF_ASSERT_OK(builder.Finalize(&op.node_def)); 51 INFER_OK(op, "?;?;?;?", "?;?"); // output rank unknown 54 INFER_OK(op, "?;?;?;?;?;?;?", "?;?;?"); // output rank unknown
|
/external/tensorflow/tensorflow/python/framework/ |
H A D | subscribe.py | 82 for op in graph.get_operations(): 83 for control_input in op.control_inputs: 86 control_outputs[control_input].add(op) 89 def get_control_outputs(self, op): 90 """Return the control outputs for a given op. 93 op: The op to fetch control outputs for. 97 if op.graph not in self.cache: 98 control_outputs = self.calc_control_outputs(op.graph) 99 self.cache[op [all...] |
/external/compiler-rt/lib/builtins/arm/ |
H A D | sync-ops.h | 18 #define SYNC_OP_4(op) \ 22 DEFINE_COMPILERRT_THUMB_FUNCTION(__sync_fetch_and_ ## op) \ 25 LOCAL_LABEL(tryatomic_ ## op): \ 27 op(r2, r0, r1) ; \ 30 bne LOCAL_LABEL(tryatomic_ ## op) ; \ 34 #define SYNC_OP_8(op) \ 38 DEFINE_COMPILERRT_THUMB_FUNCTION(__sync_fetch_and_ ## op) \ 42 LOCAL_LABEL(tryatomic_ ## op): \ 44 op(r4, r5, r0, r1, r2, r3) ; \ 47 bne LOCAL_LABEL(tryatomic_ ## op) ; \ [all...] |
/external/syslinux/core/ |
H A D | lstadjust.pl | 40 ($label, $op, $arg, $tail) = split(/\s+/, $source); 41 if ($op =~ /^(|\[)section$/i) { 43 } elsif ($op =~ /^(absolute|\[absolute)$/i) { 45 } elsif ($op =~ /^struc$/i) { 48 } elsif ($op =~ /^endstruc$/i) {
|
/external/tensorflow/tensorflow/contrib/lite/tools/ |
H A D | gen_op_registration_main.cc | 65 for (const auto& op : builtin_ops) { 66 fout << "TfLiteRegistration* Register_" << op << "();\n"; local 74 for (const auto& op : custom_ops) { 76 << ::tflite::NormalizeCustomOpName(op) << "();\n"; 84 for (const auto& op : builtin_ops) { 85 fout << " resolver->AddBuiltin(::tflite::BuiltinOperator_" << op 86 << ", ::tflite::ops::builtin::Register_" << op << "());\n"; local 88 for (const auto& op : custom_ops) { 89 fout << " resolver->AddCustom(\"" << op 91 << ::tflite::NormalizeCustomOpName(op) << "());\ [all...] |