/sysdeps/x86_64/ |
A D | addmul_1.S | 35 push %rbx 37 lea (%rdx), %rbx 38 neg %rbx 54 add $2, %rbx 58 mov (up,%rbx,8), %rax 62 L(odd): add $1, %rbx 66 mov (up,%rbx,8), %rax 78 mov (up,%rbx,8), %rax 81 mov (rp,%rbx,8), %r10 93 L(e): add $2, %rbx [all …]
|
A D | mul_1.S | 32 push %rbx 34 cfi_rel_offset (%rbx, 0) 39 mov %rbx, %r11 69 mov %rdx, %rbx 75 mov %rax, %rbx 98 adc %rdx, %rbx 102 add %rax, %rbx 106 mov %rbx, 24(rp,n,8) 108 mov %r8, %rbx # zero 124 pop %rbx [all …]
|
A D | __longjmp.S | 81 cmpq %rbx, %rax 82 cmovb %rax, %rbx 83 incsspq %rbx 84 subq %rbx, %rax 95 cfi_offset(%rbx,JB_RBX*8) 100 movq (JB_RBX*8)(%rdi),%rbx
|
A D | dl-trampoline.h | 37 # define BASE rbx 74 cfi_rel_offset(%rbx, 0) 146 movq (%rsp), %rbx 147 cfi_restore(%rbx) 179 movq %rbx, (%rsp) 205 movq %rsp, 24(%rbx) 216 lea 48(%rbx), %RAX_LP 368 movq (%rsp), %rbx 369 cfi_restore(%rbx) 481 movq (%rsp), %rbx [all …]
|
A D | setjmp.S | 32 movq %rbx, (JB_RBX*8)(%rdi)
|
/sysdeps/unix/sysv/linux/x86_64/ |
A D | ____longjmp_chk.S | 87 cfi_register (%rsi, %rbx) 138 cmpq %rbx, %rax 139 cmovb %rax, %rbx 140 incsspq %rbx 141 subq %rbx, %rax 151 cfi_offset(%rbx,JB_RBX*8) 156 movq (JB_RBX*8)(%rdi), %rbx
|
A D | __start_context.S | 105 movq %rbx, %rsp
|
A D | setcontext.S | 65 cfi_offset(%rbx,oRBX) 75 movq oRBX(%rdx), %rbx
|
/sysdeps/x86_64/fpu/multiarch/ |
A D | svml_s_acosf8_core_avx2.S | 114 # LOE rbx r12 r13 r14 r15 edx ymm0 ymm5 136 # LOE rbx r12 r13 r14 r15 edx ymm0 139 # LOE rbx r12 r13 r14 r15 eax edx 153 # LOE rbx r15 r12d r13d 164 # LOE rbx r15 r12d r13d 176 # LOE rbx r15 r12d r13d 194 # LOE rbx r12 r13 r14 r15 ymm0 204 # LOE rbx r14 r15 r12d r13d xmm0 210 # LOE rbx r15 r12d r13d
|
A D | svml_s_acosf16_core_avx512.S | 121 # LOE rbx r12 r13 r14 r15 edx zmm0 zmm4 143 # LOE rbx r12 r13 r14 r15 edx zmm0 146 # LOE rbx r12 r13 r14 r15 eax edx 160 # LOE rbx r15 r12d r13d 171 # LOE rbx r15 r12d r13d 183 # LOE rbx r15 r12d r13d 201 # LOE rbx r12 r13 r14 r15 zmm0 211 # LOE rbx r14 r15 r12d r13d xmm0 217 # LOE rbx r15 r12d r13d
|
A D | svml_s_acosf4_core_sse4.S | 131 # LOE rbx rbp r12 r13 r14 r15 edx xmm0 xmm7 151 # LOE rbx rbp r12 r13 r14 r15 edx 162 # LOE rbx rbp r15 r12d r13d 173 # LOE rbx rbp r15 r12d r13d 185 # LOE rbx rbp r15 r12d r13d 200 # LOE rbx rbp r12 r13 r14 r15 xmm7 210 # LOE rbx rbp r14 r15 r12d r13d xmm0 216 # LOE rbx rbp r15 r12d r13d
|
A D | svml_d_acos4_core_avx2.S | 126 # LOE rbx r12 r13 r14 r15 edx ymm0 ymm5 148 # LOE rbx r12 r13 r14 r15 edx ymm0 151 # LOE rbx r12 r13 r14 r15 eax edx 165 # LOE rbx r15 r12d r13d 176 # LOE rbx r15 r12d r13d 188 # LOE rbx r15 r12d r13d 206 # LOE rbx r12 r13 r14 r15 ymm0 216 # LOE rbx r14 r15 r12d r13d xmm0 222 # LOE rbx r15 r12d r13d
|
A D | svml_d_acos8_core_avx512.S | 148 # LOE rbx r12 r13 r14 r15 edx zmm0 zmm6 170 # LOE rbx r12 r13 r14 r15 edx zmm0 173 # LOE rbx r12 r13 r14 r15 eax edx 187 # LOE rbx r15 r12d r13d 198 # LOE rbx r15 r12d r13d 210 # LOE rbx r15 r12d r13d 228 # LOE rbx r12 r13 r14 r15 zmm0 238 # LOE rbx r14 r15 r12d r13d xmm0 244 # LOE rbx r15 r12d r13d
|
A D | svml_d_acos2_core_sse4.S | 156 # LOE rbx rbp r12 r13 r14 r15 edx xmm0 xmm5 175 # LOE rbx rbp r12 r13 r14 r15 edx 186 # LOE rbx rbp r15 r12d r13d 197 # LOE rbx rbp r15 r12d r13d 209 # LOE rbx rbp r15 r12d r13d 224 # LOE rbx rbp r12 r13 r14 r15 xmm0 234 # LOE rbx rbp r14 r15 r12d r13d xmm0 240 # LOE rbx rbp r15 r12d r13d
|
A D | svml_d_sincos4_core_avx2.S | 205 movq %rbx, 168(%rsp) 206 movq %rdi, %rbx 225 movq %rbx, %rdi 241 movq 168(%rsp), %rbx
|
/sysdeps/x86_64/fpu/ |
A D | svml_d_sincos2_core.S | 63 pushq %rbx 73 movq %rdi, %rbx 78 leal 8(%rbx), %edi 96 popq %rbx
|
A D | svml_s_sincosf4_core.S | 85 pushq %rbx 95 movq %rdi, %rbx 100 leal 4(%rbx), %edi 104 leal 8(%rbx), %edi 108 leal 12(%rbx), %edi 138 popq %rbx
|
A D | svml_s_wrapper_impl.h | 84 pushq %rbx 86 cfi_rel_offset (%rbx, 0) 88 movq %rsi, %rbx 101 movss %xmm0, (%rbx) 111 movss %xmm0, 4(%rbx) 122 movss %xmm0, 8(%rbx) 128 movss %xmm0, 12(%rbx) 131 popq %rbx 133 cfi_restore (%rbx)
|
A D | svml_d_sincos4_core.S | 95 pushq %rbx 97 movq %rdi, %rbx 106 leal 16(%rbx), %edi 135 popq %rbx
|
A D | svml_d_sincos4_core_avx.S | 90 pushq %rbx 92 movq %rdi, %rbx 101 leal 16(%rbx), %edi 130 popq %rbx
|
A D | svml_d_wrapper_impl.h | 62 pushq %rbx 64 cfi_rel_offset (%rbx, 0) 66 movq %rsi, %rbx 80 movsd %xmm0, (%rbx) 86 movsd %xmm0, 8(%rbx) 89 popq %rbx 91 cfi_restore (%rbx)
|
A D | svml_d_sincos8_core.S | 117 pushq %rbx 119 movq %rdi, %rbx 127 leal 32(%rbx), %edi 178 popq %rbx
|
A D | svml_s_sincosf8_core_avx.S | 117 pushq %rbx 119 movq %rdi, %rbx 130 leal 16(%rbx), %edi 185 popq %rbx
|
A D | svml_s_sincosf8_core.S | 121 pushq %rbx 123 movq %rdi, %rbx 132 leal 16(%rbx), %edi 183 popq %rbx
|
/sysdeps/unix/x86_64/ |
A D | sysdep.h | 31 #define r1 %rbx /* Secondary return-value register. */
|