@@ -52,7 +52,6 @@ TEST_F(TcCudaMapper1DReductionTest, DISABLED_Reduction1Dv0) {
52
52
.tile (0 )
53
53
.mapToBlocks ({})
54
54
.mapToThreads ({16 });
55
- LOG (INFO) << mappingOptions << endl;
56
55
at::Tensor A = at::CUDA (at::kFloat ).rand ({M});
57
56
Check (A, mappingOptions, 0 );
58
57
}
@@ -62,7 +61,6 @@ TEST_F(TcCudaMapper1DReductionTest, Reduction1Dv1) {
62
61
.tile (0 )
63
62
.mapToBlocks ({1 })
64
63
.mapToThreads ({16 });
65
- LOG (INFO) << mappingOptions << endl;
66
64
at::Tensor A = at::CUDA (at::kFloat ).rand ({M});
67
65
Check (A, mappingOptions, 1 );
68
66
}
@@ -72,7 +70,6 @@ TEST_F(TcCudaMapper1DReductionTest, Reduction1Dv2) {
72
70
.tile (0 )
73
71
.mapToBlocks ({1 })
74
72
.mapToThreads ({16 });
75
- LOG (INFO) << mappingOptions << endl;
76
73
at::Tensor A = at::CUDA (at::kFloat ).rand ({M});
77
74
Check (A, mappingOptions, 2 );
78
75
}
@@ -82,7 +79,6 @@ TEST_F(TcCudaMapper1DReductionTest, Reduction1Dv3) {
82
79
.tile (0 )
83
80
.mapToBlocks ({1 })
84
81
.mapToThreads ({16 });
85
- LOG (INFO) << mappingOptions << endl;
86
82
at::Tensor A = at::CUDA (at::kFloat ).rand ({M});
87
83
Check (A, mappingOptions, 3 );
88
84
}
@@ -97,7 +93,6 @@ TEST_F(TcCudaMapper2DReductionTest, Reduction2D1) {
97
93
.mapToBlocks ({1 , 1 })
98
94
.mapToThreads ({32 })
99
95
.matchLibraryCalls (true );
100
- LOG (INFO) << mappingOptions << endl;
101
96
at::Tensor A = at::CUDA (at::kFloat ).rand ({M, N});
102
97
Check (A, mappingOptions);
103
98
}
@@ -119,7 +114,6 @@ struct TcCudaMapper2DReductionStressTest : public TcCudaMapper2DReductionTest {
119
114
.mapToBlocks ({1 })
120
115
.mapToThreads ({tix, tiy})
121
116
.matchLibraryCalls (true );
122
- LOG (INFO) << mappingOptions << endl;
123
117
at::Tensor A = ones ? at::CUDA (at::kFloat ).ones ({M, N})
124
118
: at::CUDA (at::kFloat ).rand ({M, N});
125
119
return TcCudaMapper2DReductionTest::Check (A, mappingOptions, skipCheck);
@@ -196,8 +190,6 @@ TEST_F(TcCudaMapperMatmulTest, Matmul1DSchedule) {
196
190
.mapToBlocks ({M, N})
197
191
.mapToThreads ({std::min (32u , K)})
198
192
.matchLibraryCalls (true );
199
- LOG (INFO) << mappingOptions << endl;
200
-
201
193
at::Tensor A = at::CUDA (at::kFloat ).rand ({M, K});
202
194
at::Tensor B = at::CUDA (at::kFloat ).rand ({K, N});
203
195
Check (A, B, mappingOptions);
@@ -213,8 +205,6 @@ TEST_F(TcCudaMapperMatmulTest, Matmul1DScheduleMultipleOccurrence) {
213
205
.mapToBlocks ({8 })
214
206
.mapToThreads ({16 })
215
207
.matchLibraryCalls (true );
216
- LOG (INFO) << mappingOptions << endl;
217
-
218
208
at::Tensor A = at::CUDA (at::kFloat ).rand ({M, K});
219
209
at::Tensor B = at::CUDA (at::kFloat ).rand ({K, N});
220
210
Check (A, B, mappingOptions);
@@ -226,8 +216,6 @@ TEST_F(TcCudaMapperMatmulTest, Matmul3DSchedule) {
226
216
.mapToBlocks ({1 , 1 , 1 })
227
217
.mapToThreads ({4 , 1 , 1 });
228
218
mappingOptions.matchLibraryCalls (true );
229
- LOG (INFO) << mappingOptions << endl;
230
-
231
219
at::Tensor A = at::CUDA (at::kFloat ).rand ({M, K});
232
220
at::Tensor B = at::CUDA (at::kFloat ).rand ({K, N});
233
221
Check (A, B, mappingOptions);
@@ -239,8 +227,6 @@ TEST_F(TcCudaMapperMatmulTest, Matmul3DScheduleMultipleOccurrence) {
239
227
.mapToBlocks ({8 })
240
228
.mapToThreads ({16 })
241
229
.matchLibraryCalls (true );
242
- LOG (INFO) << mappingOptions << endl;
243
-
244
230
at::Tensor A = at::CUDA (at::kFloat ).rand ({M, K});
245
231
at::Tensor B = at::CUDA (at::kFloat ).rand ({K, N});
246
232
Check (A, B, mappingOptions);
@@ -258,8 +244,6 @@ TEST_F(TcCudaMapperBatchMatmulTest, BatchMatmul) {
258
244
.usePrivateMemory (true )
259
245
.useSharedMemory (true );
260
246
mappingOptions.matchLibraryCalls (true );
261
- LOG (INFO) << mappingOptions << endl;
262
-
263
247
at::Tensor A = at::CUDA (at::kFloat ).rand ({50 , 26 , 72 });
264
248
at::Tensor B = at::CUDA (at::kFloat ).rand ({50 , 72 , 26 });
265
249
Check (A, B, mappingOptions);
0 commit comments