mirror of
https://github.com/tensorflow/tensorflow.git
synced 2024-11-21 21:05:19 +00:00
Remove some macros
This commit is contained in:
parent
a114edefdb
commit
7e8898c42f
@ -58,7 +58,6 @@ limitations under the License.
|
||||
#include "tensorflow/stream_executor/gpu/gpu_asm_opts.h"
|
||||
#include "tensorflow/stream_executor/gpu/redzone_allocator.h"
|
||||
#include "tensorflow/stream_executor/tf_allocator_adapter.h"
|
||||
#include "third_party/gpus/cudnn/cudnn.h"
|
||||
#endif // GOOGLE_CUDA
|
||||
|
||||
namespace {
|
||||
|
@ -20,9 +20,6 @@ limitations under the License.
|
||||
#if GOOGLE_CUDA || TENSORFLOW_USE_ROCM
|
||||
#include "tensorflow/core/protobuf/autotuning.pb.h"
|
||||
#endif // GOOGLE_CUDA || TENSORFLOW_USE_ROCM
|
||||
#if GOOGLE_CUDA
|
||||
#include "third_party/gpus/cudnn/cudnn.h"
|
||||
#endif // GOOGLE_CUDA
|
||||
|
||||
namespace tensorflow {
|
||||
|
||||
|
@ -67,7 +67,6 @@ limitations under the License.
|
||||
#include "tensorflow/stream_executor/gpu/gpu_asm_opts.h"
|
||||
#include "tensorflow/stream_executor/gpu/redzone_allocator.h"
|
||||
#include "tensorflow/stream_executor/tf_allocator_adapter.h"
|
||||
#include "third_party/gpus/cudnn/cudnn.h"
|
||||
#endif // GOOGLE_CUDA
|
||||
|
||||
namespace tensorflow {
|
||||
|
@ -43,7 +43,6 @@ using stream_executor::dnn::DimIndex;
|
||||
#include "tensorflow/stream_executor/gpu/asm_compiler.h"
|
||||
#include "tensorflow/stream_executor/gpu/redzone_allocator.h"
|
||||
#include "tensorflow/stream_executor/tf_allocator_adapter.h"
|
||||
#include "third_party/gpus/cudnn/cudnn.h"
|
||||
#endif // GOOGLE_CUDA
|
||||
|
||||
namespace tensorflow {
|
||||
|
@ -708,7 +708,7 @@ class Conv2DTest(test.TestCase):
|
||||
tensor_in_sizes=[1, 2, 3, 3],
|
||||
filter_in_sizes=[2, 2, 3, 2],
|
||||
strides=[1, 1],
|
||||
padding=[[0, 3], [0, 0]], tol=3e-5)
|
||||
padding=[[0, 3], [0, 0]], tol=2e-5)
|
||||
|
||||
self._VerifyExplicitPaddings(
|
||||
tensor_in_sizes=[2, 2, 4, 3],
|
||||
|
@ -29,6 +29,7 @@ limitations under the License.
|
||||
#include "tensorflow/core/platform/macros.h"
|
||||
#include "tensorflow/core/platform/thread_annotations.h"
|
||||
#include "tensorflow/stream_executor/blas.h"
|
||||
#include "tensorflow/stream_executor/cuda/cuda_dnn.h"
|
||||
#include "tensorflow/stream_executor/device_memory.h"
|
||||
#include "tensorflow/stream_executor/dnn.h"
|
||||
#include "tensorflow/stream_executor/event.h"
|
||||
@ -41,10 +42,6 @@ limitations under the License.
|
||||
#include "tensorflow/stream_executor/stream_executor_pimpl.h"
|
||||
#include "tensorflow/stream_executor/temporary_memory_manager.h"
|
||||
|
||||
#if GOOGLE_CUDA
|
||||
#include "tensorflow/stream_executor/cuda/cuda_dnn.h"
|
||||
#endif // GOOGLE_CUDA
|
||||
|
||||
namespace stream_executor {
|
||||
|
||||
namespace host {
|
||||
@ -366,7 +363,6 @@ class Stream {
|
||||
DeviceMemory<OutputType> *output, ScratchAllocator *scratch_allocator,
|
||||
const dnn::AlgorithmConfig &plan_config,
|
||||
dnn::ProfileResult *output_profile_result) {
|
||||
#if GOOGLE_CUDA
|
||||
dnn::DnnSupport *dnn = parent_->AsDnn();
|
||||
if (dnn) {
|
||||
gpu::CudnnSupport *cudnn_dnn = dynamic_cast<gpu::CudnnSupport*>(dnn);
|
||||
@ -377,8 +373,6 @@ class Stream {
|
||||
output_descriptor, *output, convolution_descriptor, plan_config,
|
||||
scratch_allocator, output_profile_result);
|
||||
}
|
||||
#endif // GOOGLE_CUDA
|
||||
return port::UnimplementedError("DNN library is not found.");
|
||||
}
|
||||
|
||||
port::Status FusedConvolveWithAlgorithm(
|
||||
|
Loading…
Reference in New Issue
Block a user