/external/mdnsresponder/mDNSWindows/DLL.NET/ |
H A D | dnssd_NET.h | 164 m_impl = new RecordRefImpl; 165 m_impl->m_ref = NULL; 170 delete m_impl; 180 RecordRefImpl * m_impl; member in class:Apple::DNSSD::RecordRef 415 ServiceRefImpl * m_impl; member in class:Apple::DNSSD::ServiceRef 455 m_impl = new TextRecordImpl(); 456 TXTRecordCreate(&m_impl->m_ref, 0, NULL); 461 TXTRecordDeallocate(&m_impl->m_ref); 462 delete m_impl; 472 TextRecordImpl * m_impl; member in class:Apple::DNSSD::TextRecord [all...] |
/external/v8/src/inspector/ |
H A D | string-16.h | 25 : m_impl(other.m_impl), hash_code(other.hash_code) {} 27 : m_impl(std::move(other.m_impl)), hash_code(other.hash_code) {} 28 String16(const UChar* characters, size_t size) : m_impl(characters, size) {} 30 : m_impl(characters) {} 34 m_impl.resize(size); 35 for (size_t i = 0; i < size; ++i) m_impl[i] = characters[i]; 37 explicit String16(const std::basic_string<UChar>& impl) : m_impl(impl) {} 40 m_impl 111 std::basic_string<UChar> m_impl; member in class:v8_inspector::String16 [all...] |
/external/deqp/external/vulkancts/framework/vulkan/ |
H A D | vkPrograms.hpp | 88 explicit Iterator (const IteratorImpl& i) : m_impl(i) {} 90 Iterator& operator++ (void) { ++m_impl; return *this; } 93 const std::string& getName (void) const { return m_impl->first; } 94 const Program& getProgram (void) const { return *m_impl->second; } 96 bool operator== (const Iterator& other) const { return m_impl == other.m_impl; } 97 bool operator!= (const Iterator& other) const { return m_impl != other.m_impl; } 101 IteratorImpl m_impl; member in class:vk::ProgramCollection::Iterator
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
H A D | TensorArgMax.h | 96 : m_impl(op.expression(), device) { } 99 return m_impl.dimensions(); 103 m_impl.evalSubExprsIfNeeded(NULL); 107 m_impl.cleanup(); 112 return CoeffReturnType(index, m_impl.coeff(index)); 117 return m_impl.costPerCoeff(vectorized) + TensorOpCost(0, 0, 1); 123 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator 224 m_impl(op.expression().index_tuples().reduce(op.reduce_dims(), op.reduce_op()), device), 239 return m_impl.dimensions(); 243 m_impl [all...] |
H A D | TensorConversion.h | 56 : m_impl(impl) {} 60 return internal::pcast<SrcPacket, TgtPacket>(m_impl.template packet<LoadMode>(index)); 64 const TensorEvaluator& m_impl; member in struct:Eigen::PacketConverter 72 : m_impl(impl) {} 78 SrcPacket src1 = m_impl.template packet<LoadMode>(index); 79 SrcPacket src2 = m_impl.template packet<LoadMode>(index + SrcPacketSize); 85 const TensorEvaluator& m_impl; member in struct:Eigen::PacketConverter 92 : m_impl(impl) {} 98 SrcPacket src1 = m_impl.template packet<LoadMode>(index); 99 SrcPacket src2 = m_impl 107 const TensorEvaluator& m_impl; member in struct:Eigen::PacketConverter 140 const TensorEvaluator& m_impl; member in struct:Eigen::PacketConverter 274 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorEvalTo.h | 112 : m_impl(op.expression(), device), m_device(device), 125 EIGEN_DEVICE_FUNC const Dimensions& dimensions() const { return m_impl.dimensions(); } 130 return m_impl.evalSubExprsIfNeeded(m_buffer); 134 m_buffer[i] = m_impl.coeff(i); 137 internal::pstoret<CoeffReturnType, PacketReturnType, Aligned>(m_buffer + i, m_impl.template packet<TensorEvaluator<ArgType, Device>::IsAligned ? Aligned : Unaligned>(i)); 141 m_impl.cleanup(); 158 return m_impl.costPerCoeff(vectorized) + 166 const TensorEvaluator<ArgType, Device>& impl() const { return m_impl; } 171 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator
|
H A D | TensorEvaluator.h | 49 : m_data(const_cast<typename internal::traits<Derived>::template MakePointer<Scalar>::Type>(m.data())), m_dims(m.dimensions()), m_device(device), m_impl(m) 53 const Derived& derived() const { return m_impl; } 120 const Derived& m_impl; member in struct:Eigen::TensorEvaluator 169 const Derived& derived() const { return m_impl; } 172 : m_data(m.data()), m_dims(m.dimensions()), m_device(device), m_impl(m) 219 const Derived& m_impl; member in struct:Eigen::TensorEvaluator
|
H A D | TensorForcedEval.h | 111 : m_impl(op.expression(), device), m_op(op.expression()), m_device(device), m_buffer(NULL) 114 EIGEN_DEVICE_FUNC const Dimensions& dimensions() const { return m_impl.dimensions(); } 117 const Index numValues = internal::array_prod(m_impl.dimensions()); 154 const TensorEvaluator<ArgType, Device>& impl() { return m_impl; } 158 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator
|
H A D | TensorInflation.h | 99 : m_impl(op.expression(), device), m_strides(op.strides()) 101 m_dimensions = m_impl.dimensions(); 112 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 133 m_impl.evalSubExprsIfNeeded(NULL); 137 m_impl.cleanup(); 181 return m_impl.coeff(inputIndex); 207 const double input_size = m_impl.dimensions().TotalSize(); 211 return m_impl.costPerCoeff(vectorized) + 222 TensorEvaluator<ArgType, Device> m_impl; member in namespace:Eigen
|
H A D | TensorLayoutSwap.h | 127 : m_impl(op.expression(), device) 130 m_dimensions[i] = m_impl.dimensions()[NumDims-1-i]; 141 return m_impl.evalSubExprsIfNeeded(data); 144 m_impl.cleanup(); 149 return m_impl.coeff(index); 155 return m_impl.template packet<LoadMode>(index); 159 return m_impl.costPerCoeff(vectorized); 162 EIGEN_DEVICE_FUNC Scalar* data() const { return m_impl.data(); } 164 const TensorEvaluator<ArgType, Device>& impl() const { return m_impl; } 167 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorPatch.h | 102 : m_impl(op.expression(), device) 105 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 147 m_impl.evalSubExprsIfNeeded(NULL); 152 m_impl.cleanup(); 181 return m_impl.coeff(inputIndex); 233 PacketReturnType rslt = m_impl.template packet<Unaligned>(inputIndices[0]); 238 values[0] = m_impl.coeff(inputIndices[0]); 239 values[PacketSize-1] = m_impl.coeff(inputIndices[1]); 252 return m_impl.costPerCoeff(vectorized) + 264 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorReverse.h | 122 : m_impl(op.expression(), device), m_reverse(op.reverse()) 128 m_dimensions = m_impl.dimensions(); 146 m_impl.evalSubExprsIfNeeded(NULL); 150 m_impl.cleanup(); 191 return m_impl.coeff(reverseIndex(index)); 221 return m_impl.costPerCoeff(vectorized) + 230 TensorEvaluator<ArgType, Device> m_impl; member in namespace:Eigen 267 return this->m_impl.coeffRef(this->reverseIndex(index));
|
H A D | TensorShuffling.h | 120 : m_impl(op.expression(), device) 122 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 154 m_impl.evalSubExprsIfNeeded(NULL); 158 m_impl.cleanup(); 163 return m_impl.coeff(srcCoeff(index)); 184 return m_impl.costPerCoeff(vectorized) + 213 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator 245 return this->m_impl.coeffRef(this->srcCoeff(index));
|
H A D | TensorChipping.h | 153 : m_impl(op.expression(), device), m_dim(op.dim()), m_device(device) 158 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 189 m_impl.evalSubExprsIfNeeded(NULL); 194 m_impl.cleanup(); 199 return m_impl.coeff(srcCoeff(index)); 215 values[i] = m_impl.coeff(inputIndex); 224 return m_impl.template packet<LoadMode>(index + m_inputOffset); 230 return m_impl.template packet<LoadMode>(inputIndex); 262 return m_impl.costPerCoeff(vectorized) + 267 CoeffReturnType* result = const_cast<CoeffReturnType*>(m_impl 304 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorPadding.h | 104 : m_impl(op.expression(), device), m_padding(op.padding()), m_paddingValue(op.padding_value()) 112 m_dimensions = m_impl.dimensions(); 116 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 139 m_impl.evalSubExprsIfNeeded(NULL); 143 m_impl.cleanup(); 177 return m_impl.coeff(inputIndex); 190 TensorOpCost cost = m_impl.costPerCoeff(vectorized); 239 const double in = static_cast<double>(m_impl.dimensions()[i]); 309 return m_impl.template packet<Unaligned>(inputIndex); 367 return m_impl 386 TensorEvaluator<ArgType, Device> m_impl; member in namespace:Eigen [all...] |
H A D | TensorRef.h | 48 TensorLazyEvaluatorReadOnly(const Expr& expr, const Device& device) : m_impl(expr, device), m_dummy(Scalar(0)) { 49 m_dims = m_impl.dimensions(); 50 m_impl.evalSubExprsIfNeeded(NULL); 53 m_impl.cleanup(); 60 return m_impl.data(); 64 return m_impl.coeff(index); 72 TensorEvaluator<Expr, Device> m_impl; member in class:Eigen::internal::TensorLazyEvaluatorReadOnly 89 return this->m_impl.coeffRef(index);
|
H A D | TensorScan.h | 106 : m_impl(op.expression(), device), 110 m_size(m_impl.dimensions()[op.axis()]), 119 const Dimensions& dims = m_impl.dimensions(); 132 return m_impl.dimensions(); 152 return m_impl; 160 m_impl.evalSubExprsIfNeeded(NULL); 197 m_impl.cleanup(); 201 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator
|
H A D | TensorStriding.h | 120 : m_impl(op.expression(), device) 122 m_dimensions = m_impl.dimensions(); 127 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 152 m_impl.evalSubExprsIfNeeded(NULL); 156 m_impl.cleanup(); 161 return m_impl.coeff(srcCoeff(index)); 196 PacketReturnType rslt = m_impl.template packet<Unaligned>(inputIndices[0]); 201 values[0] = m_impl.coeff(inputIndices[0]); 202 values[PacketSize-1] = m_impl.coeff(inputIndices[1]); 220 return m_impl 252 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorMorphing.h | 113 : m_impl(op.expression(), device), m_dimensions(op.dimensions()) 117 eigen_assert(internal::array_prod(m_impl.dimensions()) == internal::array_prod(op.dimensions())); 128 return m_impl.evalSubExprsIfNeeded(data); 131 m_impl.cleanup(); 136 return m_impl.coeff(index); 142 return m_impl.template packet<LoadMode>(index); 146 return m_impl.costPerCoeff(vectorized); 149 EIGEN_DEVICE_FUNC Scalar* data() const { return const_cast<Scalar*>(m_impl.data()); } 151 EIGEN_DEVICE_FUNC const TensorEvaluator<ArgType, Device>& impl() const { return m_impl; } 154 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator 522 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator 841 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorFFT.h | 143 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE TensorEvaluator(const XprType& op, const Device& device) : m_fft(op.fft()), m_impl(op.expression(), device), m_data(NULL), m_device(device) { 144 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 169 m_impl.evalSubExprsIfNeeded(NULL); 185 m_impl.cleanup(); 212 buf[i] = MakeComplex<internal::is_same<InputScalar, RealScalar>::value>()(m_impl.coeff(i)); 568 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator
|
H A D | TensorImagePatch.h | 175 : m_impl(op.expression(), device) 181 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 316 m_impl.evalSubExprsIfNeeded(NULL); 321 m_impl.cleanup(); 359 return m_impl.coeff(inputIndex); 414 return m_impl.template packet<Unaligned>(inputIndex); 423 const TensorEvaluator<ArgType, Device>& impl() const { return m_impl; } 444 return m_impl.costPerCoeff(vectorized) + 503 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator
|
H A D | TensorReduction.h | 144 reducer.reduce(self.m_impl.coeff(input), accum); 151 reducer.reduce(self.m_impl.coeff(index), accum); 160 reducer.reduce(self.m_impl.coeff(firstIndex + j), &accum); 173 reducer.reducePacket(self.m_impl.template packet<Unaligned>(firstIndex + j), &p); 177 reducer.reduce(self.m_impl.coeff(firstIndex + j), &accum); 206 reducer.reducePacket(self.m_impl.template packet<Unaligned>(input), accum); 223 const typename Self::Index num_coeffs = array_prod(self.m_impl.dimensions()); 253 const Index num_coeffs = array_prod(self.m_impl.dimensions()); 259 self.m_impl.costPerCoeff(Vectorizable) + 412 : m_impl(o 757 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator [all...] |
H A D | TensorVolumePatch.h | 188 : m_impl(op.expression(), device) 194 const typename TensorEvaluator<ArgType, Device>::Dimensions& input_dims = m_impl.dimensions(); 345 m_impl.evalSubExprsIfNeeded(NULL); 350 m_impl.cleanup(); 405 return m_impl.coeff(inputIndex); 491 return m_impl.template packet<Unaligned>(inputIndex); 507 const TensorEvaluator<ArgType, Device>& impl() const { return m_impl; } 602 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::TensorEvaluator
|
/external/deqp/framework/platform/android/ |
H A D | tcuAndroidInternals.hpp | 181 android::GraphicBuffer* m_impl; member in class:tcu::Android::internal::GraphicBuffer
|
/external/tensorflow/tensorflow/core/kernels/ |
H A D | eigen_volume_patch.h | 53 : m_impl(op.expression(), device) { 59 m_impl.dimensions(); 244 m_impl.evalSubExprsIfNeeded(NULL); 248 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void cleanup() { m_impl.cleanup(); } 328 return m_impl.coeff(inputIndex); 426 return m_impl.template packet<Unaligned>(inputIndex); 442 const TensorEvaluator<ArgType, Device>& impl() const { return m_impl; } 535 return m_impl.coeff(inputCoords); 549 return m_impl.coeff(inputIndex); 631 TensorEvaluator<ArgType, Device> m_impl; member in struct:Eigen::CustomTensorEvaluator [all...] |