/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
H A D | sequential_thunk.cc | 35 tensorflow::Status SequentialThunk::ExecuteOnStream( function in class:xla::gpu::SequentialThunk 39 TF_RETURN_IF_ERROR(thunk->ExecuteOnStream(buffer_allocations, stream));
|
H A D | for_thunk.cc | 38 tensorflow::Status ForThunk::ExecuteOnStream( function in class:xla::gpu::ForThunk 44 body_thunk_sequence_->ExecuteOnStream(buffer_allocations, stream));
|
H A D | conditional_thunk.cc | 44 Status ConditionalThunk::ExecuteOnStream( function in class:xla::gpu::ConditionalThunk 62 TF_RETURN_IF_ERROR(true_thunk_.ExecuteOnStream(buffer_allocations, stream)); 65 false_thunk_.ExecuteOnStream(buffer_allocations, stream));
|
H A D | while_thunk.cc | 43 Status WhileThunk::ExecuteOnStream(const BufferAllocations& buffer_allocations, function in class:xla::gpu::WhileThunk 51 condition_thunk_sequence_->ExecuteOnStream(buffer_allocations, stream)); 69 body_thunk_sequence_->ExecuteOnStream(buffer_allocations, stream));
|
H A D | for_thunk.h | 40 tensorflow::Status ExecuteOnStream(
|
H A D | infeed_thunk.h | 46 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
|
H A D | tuple_thunk.cc | 25 tensorflow::Status TupleThunk::ExecuteOnStream( function in class:xla::gpu::TupleThunk
|
H A D | copy_thunk.cc | 32 tensorflow::Status HostToDeviceCopyThunk::ExecuteOnStream( function in class:xla::gpu::HostToDeviceCopyThunk 50 tensorflow::Status DeviceToDeviceCopyThunk::ExecuteOnStream( function in class:xla::gpu::DeviceToDeviceCopyThunk
|
H A D | copy_thunk.h | 42 tensorflow::Status ExecuteOnStream( 66 tensorflow::Status ExecuteOnStream(
|
H A D | conditional_thunk.h | 51 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
|
H A D | sequential_thunk.h | 42 tensorflow::Status ExecuteOnStream(
|
H A D | tuple_thunk.h | 48 tensorflow::Status ExecuteOnStream(
|
H A D | while_thunk.h | 49 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
|
H A D | cudnn_batchnorm_thunk.h | 62 Status ExecuteOnStream(const BufferAllocations& buffer_allocations, 92 Status ExecuteOnStream(const BufferAllocations& buffer_allocations, 125 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
|
H A D | fft_thunk.h | 76 tensorflow::Status ExecuteOnStream(
|
H A D | gemm_thunk.h | 51 tensorflow::Status ExecuteOnStream(
|
H A D | gpu_executable.h | 63 // This should be called before ExecuteOnStream. 75 // ExecuteOnStream will fail if the compute capability of the stream doesn't 77 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream( 111 // This string should be modified only before ExecuteOnStream.
|
H A D | kernel_thunk.h | 40 // kernel. It implements the ExecuteOnStream interface for GpuExecutable to 61 tensorflow::Status ExecuteOnStream( 76 // Describes how to load this kernel. ExecuteOnStream reuses this loader
|
H A D | thunk.h | 35 // Thunk provides the Initialize and ExecuteOnStream interface for GpuExecutable 79 // before calling ExecuteOnStream(stream). If it returns true, it's the 81 // calling ExecuteOnStream. 104 virtual tensorflow::Status ExecuteOnStream(
|
H A D | convolution_thunk.h | 68 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
|
H A D | infeed_thunk.cc | 33 Status InfeedThunk::ExecuteOnStream(const BufferAllocations& buffer_allocations, function in class:xla::gpu::InfeedThunk
|
/external/tensorflow/tensorflow/compiler/xla/service/interpreter/ |
H A D | executable.h | 46 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | executable.cc | 41 ExecuteOnStream(&run_options[0], arguments[0], 81 ExecuteOnStream(run_options, arguments, profile_ptr.get());
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
H A D | cpu_executable.h | 58 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(
|
H A D | parallel_cpu_executable.h | 62 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(
|