Skip to content
This repository was archived by the owner on Nov 17, 2023. It is now read-only.

Commit d086543

Browse files
committed
fix include in tensorrt-inl.h, fix bug following modification of buildCudaEngine, changer assert_allclose to assert_almost_equal
1 parent 21da738 commit d086543

File tree

3 files changed

+17
-24
lines changed

3 files changed

+17
-24
lines changed

src/operator/subgraph/tensorrt/nnvm_to_onnx-inl.h

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -140,17 +140,17 @@ void ConvertBatchNorm(NodeProto *node_proto,
140140
const nnvm::IndexedGraph &ig,
141141
const array_view<IndexedGraph::NodeEntry> &inputs);
142142

143-
void ConvertElementwiseSub(NodeProto *node_proto,
143+
void ConvertElementwiseAdd(NodeProto *node_proto,
144144
const NodeAttrs &attrs,
145145
const nnvm::IndexedGraph &ig,
146146
const array_view<IndexedGraph::NodeEntry> &inputs);
147147

148-
void ConvertElementwiseSub(NodeProto *node_proto,
148+
void ConvertElementwiseMul(NodeProto *node_proto,
149149
const NodeAttrs &attrs,
150150
const nnvm::IndexedGraph &ig,
151151
const array_view<IndexedGraph::NodeEntry> &inputs);
152152

153-
void ConvertElementwiseMul(NodeProto *node_proto,
153+
void ConvertElementwiseSub(NodeProto *node_proto,
154154
const NodeAttrs &attrs,
155155
const nnvm::IndexedGraph &ig,
156156
const array_view<IndexedGraph::NodeEntry> &inputs);
@@ -183,8 +183,8 @@ static const std::unordered_map<std::string, ConverterFunction> converter_map =
183183
{"Concat", ConvertConcatenate},
184184
{"Dropout", ConvertDropout},
185185
{"elemwise_add", ConvertElementwiseAdd},
186-
{"elemwise_sub", ConvertElementwiseSub},
187186
{"elemwise_mul", ConvertElementwiseMul},
187+
{"elemwise_sub", ConvertElementwiseSub},
188188
{"Flatten", ConvertFlatten},
189189
{"FullyConnected", ConvertFullyConnected},
190190
{"Pad", ConvertPad},

src/operator/subgraph/tensorrt/tensorrt-inl.h

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,13 @@
3434
#include <string>
3535
#include <vector>
3636

37+
#include "../../nn/activation-inl.h"
38+
#include "../../nn/batch_norm-inl.h"
39+
#include "../../nn/concat-inl.h"
40+
#include "../../nn/convolution-inl.h"
41+
#include "../../nn/dropout-inl.h"
42+
#include "../../nn/fully_connected-inl.h"
43+
#include "../../nn/pooling-inl.h"
3744
#include "../common.h"
3845
#include "../subgraph_property.h"
3946
#include "nnvm_to_onnx-inl.h"

tests/python/gpu/test_tensorrt.py

Lines changed: 6 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
from itertools import product
2121
import copy
2222

23-
from mxnet.test_utils import assert_allclose
23+
from mxnet.test_utils import assert_almost_equal
2424

2525
def check_unsupported_single_sym(sym):
2626
wrapped_sym = mx.sym.Group([mx.sym.identity(s) for s in sym])
@@ -74,15 +74,8 @@ def check_single_sym(sym, arg_params_shapes=None, aux_params_shapes=None,
7474

7575
trt_fp32_outputs = [arr.asnumpy() for arr in trt_fp32_executor.outputs]
7676
for j, (orig, fp16, fp32) in enumerate(zip(orig_outputs, trt_fp16_outputs, trt_fp32_outputs)):
77-
#abs_orig = abs(orig)
78-
#diff32 = abs(fp32 - orig)
79-
#diff16 = abs(fp16.astype('float32') - orig)
80-
#_atol32 = diff32 - rtol_fp32 * abs_orig
81-
#_atol16 = diff16 - rtol_fp16 * abs_orig
82-
#print("{}: diff32({:.2E}) | diff16({:.2E}) | atol32({:.2E}) | atol16({:.2E}) | orig.min({:.2E})".format(
83-
# j, diff32.max(), diff16.max(), _atol32.max(), _atol16.max(), abs_orig.min()))
84-
assert_allclose(fp32, orig, rtol=rtol_fp32, atol=atol_fp32)
85-
assert_allclose(fp16.astype('float32'), orig, rtol=rtol_fp16, atol=atol_fp16)
77+
assert_almost_equal(fp32, orig, rtol=rtol_fp32, atol=atol_fp32)
78+
assert_almost_equal(fp16.astype('float32'), orig, rtol=rtol_fp16, atol=atol_fp16)
8679

8780
def test_noop():
8881
data = mx.sym.Variable('data')
@@ -108,7 +101,7 @@ def test_fp16():
108101
executor.copy_params_from(arg_params, {})
109102
executor.forward(is_train=False)
110103
outputs = executor.outputs[0].asnumpy()
111-
assert_allclose(outputs, arr, rtol=0., atol=0.)
104+
assert_almost_equal(outputs, arr, rtol=0., atol=0.)
112105

113106
def test_convolution2d():
114107
data = mx.sym.Variable('data')
@@ -318,15 +311,8 @@ def check_batch_norm(sym, arg_params_shapes=None, aux_params_shapes=None,
318311
for j, (orig, fp16, fp32) in enumerate(zip(orig_outputs,
319312
trt_fp16_outputs,
320313
trt_fp32_outputs)):
321-
#abs_orig = abs(orig)
322-
#diff32 = abs(fp32 - orig)
323-
#diff16 = abs(fp16.astype('float32') - orig)
324-
#_atol32 = diff32 - rtol_fp32 * abs_orig
325-
#_atol16 = diff16 - rtol_fp16 * abs_orig
326-
#print("{}: diff32({:.2E}) | diff16({:.2E}) | atol32({:.2E}) | atol16({:.2E}) | orig.min({:.2E})".format(
327-
# j, diff32.max(), diff16.max(), _atol32.max(), _atol16.max(), abs_orig.min()))
328-
assert_allclose(fp32, orig, rtol=rtol_fp32, atol=atol_fp32)
329-
assert_allclose(fp16.astype('float32'), orig, rtol=rtol_fp16, atol=atol_fp16)
314+
assert_almost_equal(fp32, orig, rtol=rtol_fp32, atol=atol_fp32)
315+
assert_almost_equal(fp16.astype('float32'), orig, rtol=rtol_fp16, atol=atol_fp16)
330316

331317
def test_batch_norm():
332318
data = mx.sym.Variable('data')

0 commit comments

Comments
 (0)