Searched defs:n_batch (Results 1 - 9 of 9) sorted by relevance

/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/
H A Dneon_tensor_utils.h29 int n_batch, float* result,
32 vector, n_batch, result, result_stride);
49 int n_batch, float* result) {
51 batch_vector, n_batch, result);
61 int n_batch, float* result,
64 n_batch, result, result_stride);
67 void VectorBatchVectorAssign(const float* vector, int v_size, int n_batch, argument
69 PortableVectorBatchVectorAssign(vector, v_size, n_batch, batch_vector);
27 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
47 VectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument
59 BatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument
H A Dneon_tensor_utils.cc32 int n_batch, float* result,
45 for (int b = 0; b < n_batch; b++) {
163 int n_batch, float* result) {
180 for (int b = 0; b < n_batch; b++) {
278 int n_batch, float* result,
283 for (int b = 0; b < n_batch; b++) {
30 NeonMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
160 NeonVectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument
276 NeonBatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/
H A Dportable_tensor_utils.h33 int n_batch, float* result,
54 int n_batch, float* result,
63 int n_batch,
68 int n_batch, float* batch_vector);
108 int n_batch, float* result,
111 n_batch, result, result_stride);
127 int n_batch, float* result) {
129 n_batch, result);
139 int n_batch, float* result,
141 PortableBatchVectorBatchVectorDotProduct(vector1, vector2, v_size, n_batch,
106 MatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
125 VectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument
137 BatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument
145 VectorBatchVectorAssign(const float* vector, int v_size, int n_batch, float* batch_vector) argument
[all...]
H A Dportable_tensor_utils.cc33 int n_batch, float* result,
36 for (int b = 0; b < n_batch; b++) {
67 int n_batch, float* result,
72 for (int b = 0; b < n_batch; b++) {
92 int n_batch,
94 for (int b = 0; b < n_batch; b++) {
102 int n_batch, float* batch_vector) {
103 for (int b = 0; b < n_batch; b++) {
30 PortableMatrixBatchVectorMultiplyAccumulate(const float* matrix, int m_rows, int m_cols, const float* vector, int n_batch, float* result, int result_stride) argument
65 PortableBatchVectorBatchVectorDotProduct(const float* vector1, const float* vector2, int v_size, int n_batch, float* result, int result_stride) argument
89 PortableVectorBatchVectorCwiseProductAccumulate(const float* vector, int v_size, const float* batch_vector, int n_batch, float* result) argument
101 PortableVectorBatchVectorAssign(const float* vector, int v_size, int n_batch, float* batch_vector) argument
/external/tensorflow/tensorflow/contrib/lite/kernels/
H A Dlstm.cc36 // Input Tensors of size {n_batch, n_input}
233 const int n_batch = input->dims->data[0]; local
262 output_size->data[0] = n_batch;
268 output_state_size->data[0] = n_batch;
275 cell_size->data[0] = n_batch;
289 scratch_buffer_size->data[0] = n_batch;
296 scratch_buffer_size->data[0] = n_batch;
352 const int n_batch = input->dims->data[0]; local
371 forget_gate_scratch = scratch_buffer->data.f + n_cell * n_batch;
372 output_gate_scratch = scratch_buffer->data.f + 2 * n_cell * n_batch;
[all...]
H A Doptional_tensor_test.cc33 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, argument
37 : n_batch_(n_batch),
247 const int n_batch = 1; local
253 LSTMOpModel lstm(n_batch, n_input, n_cell, n_output,
259 {n_batch, n_input}, // input tensor
H A Dunidirectional_sequence_lstm.cc36 // Input Tensors of size {max_time, n_batch, n_input}
234 const int n_batch = input->dims->data[1]; local
264 output_size->data[1] = n_batch;
270 output_state_size->data[0] = n_batch;
277 cell_size->data[0] = n_batch;
291 scratch_buffer_size->data[0] = n_batch;
298 scratch_buffer_size->data[0] = n_batch;
355 const int n_batch = input->dims->data[1]; local
374 forget_gate_scratch = scratch_buffer->data.f + n_cell * n_batch;
375 output_gate_scratch = scratch_buffer->data.f + 2 * n_cell * n_batch;
[all...]
H A Dlstm_test.cc35 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, argument
39 : n_batch_(n_batch),
245 const int n_batch = 1; local
251 LSTMOpModel lstm(n_batch, n_input, n_cell, n_output,
257 {n_batch, n_input}, // input tensor
355 const int n_batch = 1; local
361 LSTMOpModel lstm(n_batch, n_input, n_cell, n_output,
367 {n_batch, n_input}, // input tensor
460 const int n_batch = 2; local
465 LSTMOpModel lstm(n_batch, n_inpu
[all...]
H A Dunidirectional_sequence_lstm_test.cc35 UnidirectionalLSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, argument
41 : n_batch_(n_batch),
251 const int n_batch = 1; local
259 n_batch, n_input, n_cell, n_output, sequence_length, /*use_cifg=*/false,
263 {sequence_length, n_batch, n_input}, // input tensor
359 const int n_batch = 1; local
367 n_batch, n_input, n_cell, n_output, sequence_length, /*use_cifg=*/true,
371 {sequence_length, n_batch, n_input}, // input tensor
461 const int n_batch = 2; local
468 n_batch, n_inpu
[all...]

Completed in 166 milliseconds