movaps %xmm5, %xmm7
shrl $2, %ecx # n_2_complex_blocks / 4
- je .cleanup # if zero, take short path
+ je .Lcleanup # if zero, take short path
# finish setup and loop priming
# At this point, xmm4 contains 2x2 partial sums. We need
# to compute a "horizontal complex add" across xmm4.
-.cleanup: # xmm4 = r1 i2 r3 i4
+.Lcleanup: # xmm4 = r1 i2 r3 i4
movl 20(%ebp), %eax # @result
movhlps %xmm4, %xmm0 # xmm0 = ?? ?? r1 r2
addps %xmm4, %xmm0 # xmm0 = ?? ?? r1+r3 i2+i4
shr $2, %rdx # n_2_complex_blocks / 4
movaps %xmm5, %xmm7
- je .cleanup # if zero, take short path
+ je .Lcleanup # if zero, take short path
# finish setup and loop priming
# At this point, xmm4 contains 2x2 partial sums. We need
# to compute a "horizontal complex add" across xmm4.
-.cleanup: # xmm4 = r1 i2 r3 i4
+.Lcleanup: # xmm4 = r1 i2 r3 i4
movhlps %xmm4, %xmm0 # xmm0 = ?? ?? r1 r2
addps %xmm4, %xmm0 # xmm0 = ?? ?? r1+r3 i2+i4
movlps %xmm0, (%rcx) # store low 2x32 bits (complex) to memory
movaps %xmm5, %xmm7
shrl $2, %ecx # n_2_complex_blocks / 4
- je .cleanup # if zero, take short path
+ je .Lcleanup # if zero, take short path
# finish setup and loop priming
# At this point, xmm4 contains 2x2 partial sums. We need
# to compute a "horizontal complex add" across xmm4.
-.cleanup: # xmm4 = r1 i2 r3 i4
+.Lcleanup: # xmm4 = r1 i2 r3 i4
movl 20(%ebp), %eax # @result
movhlps %xmm4, %xmm0 # xmm0 = ?? ?? r1 r2
addps %xmm4, %xmm0 # xmm0 = ?? ?? r1+r3 i2+i4
movaps %xmm5, %xmm7
shr $2, %rdx # n_2_complex_blocks / 4
- je .cleanup # if zero, take short path
+ je .Lcleanup # if zero, take short path
# finish setup and loop priming
# At this point, xmm4 contains 2x2 partial sums. We need
# to compute a "horizontal complex add" across xmm4.
-.cleanup: # xmm4 = r1 i2 r3 i4
+.Lcleanup: # xmm4 = r1 i2 r3 i4
movhlps %xmm4, %xmm0 # xmm0 = ?? ?? r1 r2
addps %xmm4, %xmm0 # xmm0 = ?? ?? r1+r3 i2+i4
movlps %xmm0, (%rcx) # store low 2x32 bits (complex) to memory
movaps %xmm5, %xmm7
shrl $2, %ecx # n_4_float_blocks / 4
- je .cleanup # if zero, take short path
+ je .Lcleanup # if zero, take short path
# finish setup and loop priming
# to compute a "horizontal add" across xmm4.
# This is a fairly nasty operation...
-.cleanup: # xmm4 = d1 d2 d3 d4
+.Lcleanup: # xmm4 = d1 d2 d3 d4
xorps %xmm0, %xmm0 # xmm0 = 0 0 0 0 (may be unnecessary)
movhlps %xmm4, %xmm0 # xmm0 = 0 0 d1 d2
addps %xmm4, %xmm0 # xmm0 = d1 d2 d1+d3 d2+d4
movaps %xmm5, %xmm7
shr $2, %rdx # n_4_float_blocks / 4
- je .cleanup # if zero, take short path
+ je .Lcleanup # if zero, take short path
# finish setup and loop priming
# to compute a "horizontal add" across xmm4.
# This is a fairly nasty operation...
-.cleanup: # xmm4 = d1 d2 d3 d4
+.Lcleanup: # xmm4 = d1 d2 d3 d4
xorps %xmm0, %xmm0 # xmm0 = 0 0 0 0 (may be unnecessary)
movhlps %xmm4, %xmm0 # xmm0 = 0 0 d1 d2
addps %xmm4, %xmm0 # xmm0 = d1 d2 d1+d3 d2+d4