- Eliminate OutfeedConfig since its not needed anymore and the input tuple shape is lost during HLO -> LMHLO MLIR conversion. PiperOrigin-RevId: 351474802 Change-Id: I1dee2a4b581698d394f5681e0fea22c7e27e98ab
50 lines
1.9 KiB
C++
50 lines
1.9 KiB
C++
/* Copyright 2018 The TensorFlow Authors. All Rights Reserved.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
==============================================================================*/
|
|
|
|
#ifndef TENSORFLOW_COMPILER_XLA_SERVICE_GPU_OUTFEED_THUNK_H_
|
|
#define TENSORFLOW_COMPILER_XLA_SERVICE_GPU_OUTFEED_THUNK_H_
|
|
|
|
#include "tensorflow/compiler/xla/service/gpu/buffer_allocations.h"
|
|
#include "tensorflow/compiler/xla/service/gpu/hlo_execution_profiler.h"
|
|
#include "tensorflow/compiler/xla/service/gpu/thunk.h"
|
|
#include "tensorflow/compiler/xla/service/hlo_instruction.h"
|
|
#include "tensorflow/core/platform/stream_executor_no_cuda.h"
|
|
|
|
namespace xla {
|
|
namespace gpu {
|
|
|
|
// A thunk that outfeeds data. Data must be already resident on the host. This
|
|
// thunk performs a device to host copy from the buffer allocated for the
|
|
// outfeed op to the host location.
|
|
class OutfeedThunk : public Thunk {
|
|
public:
|
|
// Constructs a OutfeedThunk that copies data to the host-side
|
|
// outfeed queue from the buffers in the given shape tree.
|
|
OutfeedThunk(ThunkInfo thunk_info, std::vector<ShapedSlice> source_slices);
|
|
|
|
OutfeedThunk(const OutfeedThunk&) = delete;
|
|
OutfeedThunk& operator=(const OutfeedThunk&) = delete;
|
|
|
|
Status ExecuteOnStream(const ExecuteParams& params) override;
|
|
|
|
private:
|
|
const std::vector<ShapedSlice> source_slices_;
|
|
};
|
|
|
|
} // namespace gpu
|
|
} // namespace xla
|
|
|
|
#endif // TENSORFLOW_COMPILER_XLA_SERVICE_GPU_OUTFEED_THUNK_H_
|