@@ -41,7 +41,7 @@ class TEST_EventMgrHelper {
4141 }
4242
4343 void QueueTensors (perftools::gputools::Stream* stream,
44- std::vector<Tensor> * tensors) {
44+ EventMgr::TensorReferenceVector * tensors) {
4545 mutex_lock l (em_->mu_ );
4646 em_->QueueTensors (stream, tensors);
4747 }
@@ -77,12 +77,12 @@ TEST(EventMgr, DelayedPolling) {
7777 EventMgr em (stream_exec);
7878 TEST_EventMgrHelper th (&em);
7979 EXPECT_EQ (0 , th.queue_size ());
80- std::vector<Tensor> * v = nullptr ;
80+ EventMgr::TensorReferenceVector * v = nullptr ;
8181 std::unique_ptr<gpu::Stream> stream (new gpu::Stream (stream_exec));
8282 CHECK (stream.get ());
8383 stream->Init ();
8484 for (int i = 0 ; i < 5 ; ++i) {
85- v = new std::vector<Tensor> ;
85+ v = new EventMgr::TensorReferenceVector ;
8686 th.QueueTensors (stream.get (), v);
8787 EXPECT_EQ (i + 1 , th.queue_size ());
8888 EXPECT_EQ (0 , th.free_size ());
@@ -92,7 +92,7 @@ TEST(EventMgr, DelayedPolling) {
9292 EXPECT_EQ (5 , th.free_size ());
9393 for (int j = 0 ; j < 2 ; ++j) {
9494 for (int i = 0 ; i < 5 ; ++i) {
95- v = new std::vector<Tensor> ;
95+ v = new EventMgr::TensorReferenceVector ;
9696 th.QueueTensors (stream.get (), v);
9797 EXPECT_EQ (i + 1 , th.queue_size ());
9898 EXPECT_EQ (4 - i, th.free_size ());
@@ -110,12 +110,12 @@ TEST(EventMgr, ImmediatePolling) {
110110 TEST_EventMgrHelper th (&em);
111111 EXPECT_EQ (0 , th.queue_size ());
112112 EXPECT_EQ (0 , th.free_size ());
113- std::vector<Tensor> * v = nullptr ;
113+ EventMgr::TensorReferenceVector * v = nullptr ;
114114 std::unique_ptr<gpu::Stream> stream (new gpu::Stream (stream_exec));
115115 CHECK (stream.get ());
116116 stream->Init ();
117117 for (int i = 0 ; i < 5 ; ++i) {
118- v = new std::vector<Tensor> ;
118+ v = new EventMgr::TensorReferenceVector ;
119119 em.ThenDeleteTensors (stream.get (), v);
120120 EXPECT_EQ (0 , th.queue_size ());
121121 EXPECT_EQ (1 , th.free_size ());
@@ -130,12 +130,12 @@ TEST(EventMgr, LongDelayedPolling) {
130130 TEST_EventMgrHelper th (&em);
131131 EXPECT_EQ (0 , th.queue_size ());
132132 EXPECT_EQ (0 , th.free_size ());
133- std::vector<Tensor> * v = nullptr ;
133+ EventMgr::TensorReferenceVector * v = nullptr ;
134134 std::unique_ptr<gpu::Stream> stream (new gpu::Stream (stream_exec));
135135 CHECK (stream.get ());
136136 stream->Init ();
137137 for (int i = 0 ; i < 5 ; ++i) {
138- v = new std::vector<Tensor> ;
138+ v = new EventMgr::TensorReferenceVector ;
139139 th.QueueTensors (stream.get (), v);
140140 EXPECT_EQ (1 + i, th.queue_size ());
141141 EXPECT_EQ (0 , th.free_size ());
@@ -153,12 +153,12 @@ TEST(EventMgr, NonEmptyShutdown) {
153153 TEST_EventMgrHelper th (&em);
154154 EXPECT_EQ (0 , th.queue_size ());
155155 EXPECT_EQ (0 , th.free_size ());
156- std::vector<Tensor> * v = nullptr ;
156+ EventMgr::TensorReferenceVector * v = nullptr ;
157157 std::unique_ptr<gpu::Stream> stream (new gpu::Stream (stream_exec));
158158 CHECK (stream.get ());
159159 stream->Init ();
160160 for (int i = 0 ; i < 5 ; ++i) {
161- v = new std::vector<Tensor> ;
161+ v = new EventMgr::TensorReferenceVector ;
162162 th.QueueTensors (stream.get (), v);
163163 EXPECT_EQ (1 + i, th.queue_size ());
164164 EXPECT_EQ (0 , th.free_size ());
0 commit comments