@@ -99,7 +99,7 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
99
99
b .L113 //alpha_r != 0.0 && alpha_i == 0.0
100
100
101
101
.L14:
102
- bceqz $fcc1, .L112 //alpha_r == 0.0 && alpha_i != 0.0
102
+ bceqz $fcc1, .L114 //alpha_r == 0.0 && alpha_i != 0.0
103
103
b .L111 //alpha_r == 0.0 && alpha_i == 0.0
104
104
.align 3
105
105
@@ -117,38 +117,6 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
117
117
b .L997
118
118
.align 3
119
119
120
- .L112: //alpha_r == 0.0 && alpha_i != 0.0
121
- xvld VX0, X, 0 * SIZE
122
- #ifdef DOUBLE
123
- xvld VX1, X, 4 * SIZE
124
- xvpickev.d x1, VX1, VX0
125
- xvpickod.d x2, VX1, VX0
126
- xvfmul.d x3, VXAI, x2
127
- xvfsub.d x3, VXZ, x3
128
- xvfmul.d x4, VXAI, x1
129
- xvilvl.d VX2, x4 ,x3
130
- xvilvh.d VX3, x4, x3
131
- xvst VX2, X, 0 * SIZE
132
- xvst VX3, X, 4 * SIZE
133
- addi.d X, X, 8 * SIZE
134
- #else
135
- xvld VX1, X, 8 * SIZE
136
- xvpickev.w x1, VX1, VX0
137
- xvpickod.w x2, VX1, VX0
138
- xvfmul.s x3, VXAI, x2
139
- xvfsub.s x3, VXZ, x3
140
- xvfmul.s x4, VXAI, x1
141
- xvilvl.w VX2, x4 ,x3
142
- xvilvh.w VX3, x4, x3
143
- xvst VX2, X, 0 * SIZE
144
- xvst VX3, X, 8 * SIZE
145
- addi.d X, X, 16 * SIZE
146
- #endif
147
- addi.d I, I, -1
148
- blt $r0, I, .L112
149
- b .L997
150
- .align 3
151
-
152
120
.L113: //alpha_r != 0.0 && alpha_i == 0.0
153
121
xvld VX0, X, 0 * SIZE
154
122
#ifdef DOUBLE
@@ -227,7 +195,7 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
227
195
b .L223 //alpha_r != 0.0 && alpha_i == 0.0
228
196
229
197
.L24:
230
- bceqz $fcc1, .L222 //alpha_r == 0.0 && alpha_i != 0.0
198
+ bceqz $fcc1, .L224 //alpha_r == 0.0 && alpha_i != 0.0
231
199
b .L221 //alpha_r == 0.0 && alpha_i == 0.0
232
200
.align 3
233
201
@@ -275,119 +243,6 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
275
243
b .L997
276
244
.align 3
277
245
278
- .L222: //alpha_r == 0.0 && alpha_i != 0.0
279
- #ifdef DOUBLE
280
- ld.d t1, X, 0 * SIZE
281
- ld.d t2, X, 1 * SIZE
282
- add .d X, X, INCX
283
- ld.d t3, X, 0 * SIZE
284
- ld.d t4, X, 1 * SIZE
285
- add .d X, X, INCX
286
- xvinsgr2vr.d x1, t1, 0
287
- xvinsgr2vr.d x2, t2, 0
288
- xvinsgr2vr.d x1, t3, 1
289
- xvinsgr2vr.d x2, t4, 1
290
- ld.d t1, X, 0 * SIZE
291
- ld.d t2, X, 1 * SIZE
292
- add .d X, X, INCX
293
- ld.d t3, X, 0 * SIZE
294
- ld.d t4, X, 1 * SIZE
295
- xvinsgr2vr.d x1, t1, 2
296
- xvinsgr2vr.d x2, t2, 2
297
- xvinsgr2vr.d x1, t3, 3
298
- xvinsgr2vr.d x2, t4, 3
299
- add .d X, X, INCX
300
-
301
- xvfmul.d x3, VXAI, x2
302
- xvfsub.d x3, VXZ, x3
303
- xvfmul.d x4, VXAI, x1
304
- addi.d I, I, -1
305
- xvstelm.d x3, XX, 0 * SIZE, 0
306
- xvstelm.d x4, XX, 1 * SIZE, 0
307
- add .d XX, XX, INCX
308
- xvstelm.d x3, XX, 0 * SIZE, 1
309
- xvstelm.d x4, XX, 1 * SIZE, 1
310
- add .d XX, XX, INCX
311
- xvstelm.d x3, XX, 0 * SIZE, 2
312
- xvstelm.d x4, XX, 1 * SIZE, 2
313
- add .d XX, XX, INCX
314
- xvstelm.d x3, XX, 0 * SIZE, 3
315
- xvstelm.d x4, XX, 1 * SIZE, 3
316
- #else
317
- ld.w t1, X, 0 * SIZE
318
- ld.w t2, X, 1 * SIZE
319
- add .d X, X, INCX
320
- ld.w t3, X, 0 * SIZE
321
- ld.w t4, X, 1 * SIZE
322
- add .d X, X, INCX
323
- xvinsgr2vr.w x1, t1, 0
324
- xvinsgr2vr.w x2, t2, 0
325
- xvinsgr2vr.w x1, t3, 1
326
- xvinsgr2vr.w x2, t4, 1
327
- ld.w t1, X, 0 * SIZE
328
- ld.w t2, X, 1 * SIZE
329
- add .d X, X, INCX
330
- ld.w t3, X, 0 * SIZE
331
- ld.w t4, X, 1 * SIZE
332
- xvinsgr2vr.w x1, t1, 2
333
- xvinsgr2vr.w x2, t2, 2
334
- xvinsgr2vr.w x1, t3, 3
335
- xvinsgr2vr.w x2, t4, 3
336
- add .d X, X, INCX
337
- ld.w t1, X, 0 * SIZE
338
- ld.w t2, X, 1 * SIZE
339
- add .d X, X, INCX
340
- ld.w t3, X, 0 * SIZE
341
- ld.w t4, X, 1 * SIZE
342
- add .d X, X, INCX
343
- xvinsgr2vr.w x1, t1, 4
344
- xvinsgr2vr.w x2, t2, 4
345
- xvinsgr2vr.w x1, t3, 5
346
- xvinsgr2vr.w x2, t4, 5
347
- ld.w t1, X, 0 * SIZE
348
- ld.w t2, X, 1 * SIZE
349
- add .d X, X, INCX
350
- ld.w t3, X, 0 * SIZE
351
- ld.w t4, X, 1 * SIZE
352
- xvinsgr2vr.w x1, t1, 6
353
- xvinsgr2vr.w x2, t2, 6
354
- xvinsgr2vr.w x1, t3, 7
355
- xvinsgr2vr.w x2, t4, 7
356
- add .d X, X, INCX
357
-
358
- xvfmul.s x3, VXAI, x2
359
- xvfsub.s x3, VXZ, x3
360
- xvfmul.s x4, VXAI, x1
361
- addi.d I, I, -1
362
- xvstelm.w x3, XX, 0 * SIZE, 0
363
- xvstelm.w x4, XX, 1 * SIZE, 0
364
- add .d XX, XX, INCX
365
- xvstelm.w x3, XX, 0 * SIZE, 1
366
- xvstelm.w x4, XX, 1 * SIZE, 1
367
- add .d XX, XX, INCX
368
- xvstelm.w x3, XX, 0 * SIZE, 2
369
- xvstelm.w x4, XX, 1 * SIZE, 2
370
- add .d XX, XX, INCX
371
- xvstelm.w x3, XX, 0 * SIZE, 3
372
- xvstelm.w x4, XX, 1 * SIZE, 3
373
- add .d XX, XX, INCX
374
- xvstelm.w x3, XX, 0 * SIZE, 4
375
- xvstelm.w x4, XX, 1 * SIZE, 4
376
- add .d XX, XX, INCX
377
- xvstelm.w x3, XX, 0 * SIZE, 5
378
- xvstelm.w x4, XX, 1 * SIZE, 5
379
- add .d XX, XX, INCX
380
- xvstelm.w x3, XX, 0 * SIZE, 6
381
- xvstelm.w x4, XX, 1 * SIZE, 6
382
- add .d XX, XX, INCX
383
- xvstelm.w x3, XX, 0 * SIZE, 7
384
- xvstelm.w x4, XX, 1 * SIZE, 7
385
- #endif
386
- add .d XX, XX, INCX
387
- blt $r0, I, .L222
388
- b .L997
389
- .align 3
390
-
391
246
.L223: //alpha_r != 0.0 && alpha_i == 0.0
392
247
#ifdef DOUBLE
393
248
ld.d t1, X, 0 * SIZE
0 commit comments