/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/ |
H A D | neon_tensor_utils.h | 29 int n_batch, float* result, 32 vector, n_batch, result, result_stride); 49 int n_batch, float* result) { 51 batch_vector, n_batch, result); 61 int n_batch, float* result, 64 n_batch, result, result_stride); 67 void VectorBatchVectorAssign(const float* vector, int v_size, int n_batch, argument 69 PortableVectorBatchVectorAssign(vector, v_size, n_batch, batch_vector); 27 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument 47 VectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument 59 BatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument
|
H A D | neon_tensor_utils.cc | 32 int n_batch, float* result, 45 for (int b = 0; b < n_batch; b++) { 163 int n_batch, float* result) { 180 for (int b = 0; b < n_batch; b++) { 278 int n_batch, float* result, 283 for (int b = 0; b < n_batch; b++) { 30 NeonMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument 160 NeonVectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument 276 NeonBatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/ |
H A D | portable_tensor_utils.h | 33 int n_batch, float* result, 54 int n_batch, float* result, 63 int n_batch, 68 int n_batch, float* batch_vector); 108 int n_batch, float* result, 111 n_batch, result, result_stride); 127 int n_batch, float* result) { 129 n_batch, result); 139 int n_batch, float* result, 141 PortableBatchVectorBatchVectorDotProduct(vector1, vector2, v_size, n_batch, 106 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument 125 VectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument 137 BatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument 145 VectorBatchVectorAssign(const float* vector, int v_size, int n_batch, float* batch_vector) argument [all...] |
H A D | portable_tensor_utils.cc | 33 int n_batch, float* result, 36 for (int b = 0; b < n_batch; b++) { 67 int n_batch, float* result, 72 for (int b = 0; b < n_batch; b++) { 92 int n_batch, 94 for (int b = 0; b < n_batch; b++) { 102 int n_batch, float* batch_vector) { 103 for (int b = 0; b < n_batch; b++) { 30 PortableMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument 65 PortableBatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument 89 PortableVectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument 101 PortableVectorBatchVectorAssign(const float* vector, int v_size, int n_batch, float* batch_vector) argument
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
H A D | lstm.cc | 36 // Input Tensors of size {n_batch, n_input} 233 const int n_batch = input->dims->data[0]; local 262 output_size->data[0] = n_batch; 268 output_state_size->data[0] = n_batch; 275 cell_size->data[0] = n_batch; 289 scratch_buffer_size->data[0] = n_batch; 296 scratch_buffer_size->data[0] = n_batch; 352 const int n_batch = input->dims->data[0]; local 371 forget_gate_scratch = scratch_buffer->data.f + n_cell * n_batch; 372 output_gate_scratch = scratch_buffer->data.f + 2 * n_cell * n_batch; [all...] |
H A D | optional_tensor_test.cc | 33 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, argument 37 : n_batch_(n_batch), 247 const int n_batch = 1; local 253 LSTMOpModel lstm(n_batch, n_input, n_cell, n_output, 259 {n_batch, n_input}, // input tensor
|
H A D | unidirectional_sequence_lstm.cc | 36 // Input Tensors of size {max_time, n_batch, n_input} 234 const int n_batch = input->dims->data[1]; local 264 output_size->data[1] = n_batch; 270 output_state_size->data[0] = n_batch; 277 cell_size->data[0] = n_batch; 291 scratch_buffer_size->data[0] = n_batch; 298 scratch_buffer_size->data[0] = n_batch; 355 const int n_batch = input->dims->data[1]; local 374 forget_gate_scratch = scratch_buffer->data.f + n_cell * n_batch; 375 output_gate_scratch = scratch_buffer->data.f + 2 * n_cell * n_batch; [all...] |
H A D | lstm_test.cc | 35 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, argument 39 : n_batch_(n_batch), 245 const int n_batch = 1; local 251 LSTMOpModel lstm(n_batch, n_input, n_cell, n_output, 257 {n_batch, n_input}, // input tensor 355 const int n_batch = 1; local 361 LSTMOpModel lstm(n_batch, n_input, n_cell, n_output, 367 {n_batch, n_input}, // input tensor 460 const int n_batch = 2; local 465 LSTMOpModel lstm(n_batch, n_inpu [all...] |
H A D | unidirectional_sequence_lstm_test.cc | 35 UnidirectionalLSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, argument 41 : n_batch_(n_batch), 251 const int n_batch = 1; local 259 n_batch, n_input, n_cell, n_output, sequence_length, /*use_cifg=*/false, 263 {sequence_length, n_batch, n_input}, // input tensor 359 const int n_batch = 1; local 367 n_batch, n_input, n_cell, n_output, sequence_length, /*use_cifg=*/true, 371 {sequence_length, n_batch, n_input}, // input tensor 461 const int n_batch = 2; local 468 n_batch, n_inpu [all...] |