Searched refs:ExecuteOnStream (Results 1 - 25 of 38) sorted by relevance

12

/external/tensorflow/tensorflow/compiler/xla/service/gpu/
H A Dsequential_thunk.cc35 tensorflow::Status SequentialThunk::ExecuteOnStream( function in class:xla::gpu::SequentialThunk
39 TF_RETURN_IF_ERROR(thunk->ExecuteOnStream(buffer_allocations, stream));
H A Dfor_thunk.cc38 tensorflow::Status ForThunk::ExecuteOnStream( function in class:xla::gpu::ForThunk
44 body_thunk_sequence_->ExecuteOnStream(buffer_allocations, stream));
H A Dconditional_thunk.cc44 Status ConditionalThunk::ExecuteOnStream( function in class:xla::gpu::ConditionalThunk
62 TF_RETURN_IF_ERROR(true_thunk_.ExecuteOnStream(buffer_allocations, stream));
65 false_thunk_.ExecuteOnStream(buffer_allocations, stream));
H A Dwhile_thunk.cc43 Status WhileThunk::ExecuteOnStream(const BufferAllocations& buffer_allocations, function in class:xla::gpu::WhileThunk
51 condition_thunk_sequence_->ExecuteOnStream(buffer_allocations, stream));
69 body_thunk_sequence_->ExecuteOnStream(buffer_allocations, stream));
H A Dfor_thunk.h40 tensorflow::Status ExecuteOnStream(
H A Dinfeed_thunk.h46 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
H A Dtuple_thunk.cc25 tensorflow::Status TupleThunk::ExecuteOnStream( function in class:xla::gpu::TupleThunk
H A Dcopy_thunk.cc32 tensorflow::Status HostToDeviceCopyThunk::ExecuteOnStream( function in class:xla::gpu::HostToDeviceCopyThunk
50 tensorflow::Status DeviceToDeviceCopyThunk::ExecuteOnStream( function in class:xla::gpu::DeviceToDeviceCopyThunk
H A Dcopy_thunk.h42 tensorflow::Status ExecuteOnStream(
66 tensorflow::Status ExecuteOnStream(
H A Dconditional_thunk.h51 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
H A Dsequential_thunk.h42 tensorflow::Status ExecuteOnStream(
H A Dtuple_thunk.h48 tensorflow::Status ExecuteOnStream(
H A Dwhile_thunk.h49 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
H A Dcudnn_batchnorm_thunk.h62 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
92 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
125 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
H A Dfft_thunk.h76 tensorflow::Status ExecuteOnStream(
H A Dgemm_thunk.h51 tensorflow::Status ExecuteOnStream(
H A Dgpu_executable.h63 // This should be called before ExecuteOnStream.
75 // ExecuteOnStream will fail if the compute capability of the stream doesn't
77 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(
111 // This string should be modified only before ExecuteOnStream.
H A Dkernel_thunk.h40 // kernel. It implements the ExecuteOnStream interface for GpuExecutable to
61 tensorflow::Status ExecuteOnStream(
76 // Describes how to load this kernel. ExecuteOnStream reuses this loader
H A Dthunk.h35 // Thunk provides the Initialize and ExecuteOnStream interface for GpuExecutable
79 // before calling ExecuteOnStream(stream). If it returns true, it's the
81 // calling ExecuteOnStream.
104 virtual tensorflow::Status ExecuteOnStream(
H A Dconvolution_thunk.h68 Status ExecuteOnStream(const BufferAllocations& buffer_allocations,
H A Dinfeed_thunk.cc33 Status InfeedThunk::ExecuteOnStream(const BufferAllocations& buffer_allocations, function in class:xla::gpu::InfeedThunk
/external/tensorflow/tensorflow/compiler/xla/service/interpreter/
H A Dexecutable.h46 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(
/external/tensorflow/tensorflow/compiler/xla/service/
H A Dexecutable.cc41 ExecuteOnStream(&run_options[0], arguments[0],
81 ExecuteOnStream(run_options, arguments, profile_ptr.get());
/external/tensorflow/tensorflow/compiler/xla/service/cpu/
H A Dcpu_executable.h58 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(
H A Dparallel_cpu_executable.h62 StatusOr<std::unique_ptr<ShapedBuffer>> ExecuteOnStream(

Completed in 182 milliseconds

12