/sysdeps/unix/sysv/linux/x86_64/ |
A D | setcontext.S | 50 popq %rdx 59 ldmxcsr oMXCSR(%rdx) 64 cfi_def_cfa(%rdx, 0) 74 movq oRSP(%rdx), %rsp 75 movq oRBX(%rdx), %rbx 146 movq oR8(%rdx), %r8 147 movq oR9(%rdx), %r9 153 movq oRDX(%rdx), %rdx 181 movq oR8(%rdx), %r8 182 movq oR9(%rdx), %r9 [all …]
|
A D | swapcontext.S | 48 movq %rdx, oRDX(%rdi) 86 movq %r12, %rdx 92 ldmxcsr oMXCSR(%rdx) 95 movq oRSP(%rdx), %rsp 96 movq oRBX(%rdx), %rbx 203 movq oR8(%rdx), %r8 204 movq oR9(%rdx), %r9 210 movq oRDX(%rdx), %rdx 239 movq oR8(%rdx), %r8 240 movq oR9(%rdx), %r9 [all …]
|
/sysdeps/x86_64/ |
A D | strcat.S | 52 movq %r8, %rdx /* magic value */ 66 movq %r8, %rdx /* magic value */ 80 movq %r8, %rdx /* magic value */ 94 movq %r8, %rdx /* magic value */ 143 movq %rax, %rdx /* move around */ 152 movb %al, (%rdx) /* Store it */ 155 incq %rdx 180 addq $8, %rdx /* Adjust pointer. */ 198 addq $8, %rdx /* Adjust pointer. */ 241 movb %al, (%rdx) /* 1st byte. */ [all …]
|
A D | wcschr.S | 40 or %rax, %rdx 51 or %rax, %rdx 65 sar %cl, %rdx 71 test %rdx, %rdx 73 bsf %rdx, %rdx 84 test %rdx, %rdx 99 or %rax, %rdx 108 or %rax, %rdx 117 or %rax, %rdx 126 or %rax, %rdx [all …]
|
A D | strcpy.S | 30 movq %rdi, %rdx /* Duplicate destination pointer. */ 40 movb %al, (%rdx) /* Store it */ 43 incq %rdx 71 addq $8, %rdx /* Adjust pointer. */ 89 addq $8, %rdx /* Adjust pointer. */ 107 addq $8, %rdx /* Adjust pointer. */ 125 addq $8, %rdx /* Adjust pointer. */ 134 movb %al, (%rdx) /* 1st byte. */ 137 incq %rdx /* Increment destination. */ 138 movb %ah, (%rdx) /* 2nd byte. */ [all …]
|
A D | strrchr.S | 39 testq %rdx, %rdx 42 xorq %rdx, %rax 62 salq $16, %rdx 69 orq %rdx, %rax 72 salq $48, %rdx 79 orq %rdx, %rax 102 testq %rdx, %rdx 133 orq %r9, %rdx 135 orq %r8, %rdx 200 orq %r8, %rdx [all …]
|
A D | addmul_1.S | 25 #define n %rdx 37 lea (%rdx), %rbx 44 lea (up,%rdx,8), up 52 lea (%rdx), %rbp 59 lea (%rdx), %r9 67 lea (%rdx), %r9 71 lea (%rdx), %rbp 82 lea (%rdx), %r9 91 lea (%rdx), %rbp 105 adc $0, %rdx [all …]
|
/sysdeps/x86_64/multiarch/ |
A D | wcscpy-ssse3.S | 55 lea 16(%rdx), %rdx 151 lea 64(%rdx), %rdx 212 lea 16(%rdx), %rdx 225 lea 16(%rdx), %rdx 238 lea 16(%rdx), %rdx 248 lea 16(%rdx), %rdx 285 lea 64(%rdx), %rdx 311 lea 16(%rdx), %rdx 324 lea 16(%rdx), %rdx 337 lea 16(%rdx), %rdx [all …]
|
A D | strcpy-ssse3.S | 115 lea 16(%rdx), %rdx 268 lea 64(%rdx), %rdx 340 lea 16(%rdx), %rdx 356 lea 16(%rdx), %rdx 372 lea 16(%rdx), %rdx 385 lea 16(%rdx), %rdx 428 lea 64(%rdx), %rdx 457 lea 16(%rdx), %rdx 473 lea 16(%rdx), %rdx 2723 xor %rdx, %rdx [all …]
|
A D | strcmp-sse42.S | 325 lea 16(%rdx), %rdx 340 lea 16(%rdx), %rdx 421 add $16, %rdx 439 add $16, %rdx 509 add $16, %rdx 527 add $16, %rdx 597 add $16, %rdx 615 add $16, %rdx 686 add $16, %rdx 704 add $16, %rdx [all …]
|
A D | strcmp-avx2.S | 116 cmpq %r11, %rdx 144 cmpq %r11, %rdx 177 cmpq %r11, %rdx 210 cmpq %r11, %rdx 265 subq %rdi, %rdx 274 subq %rdx, %r11 277 addq %rsi, %rdx 278 movq %rdx, %rsi 492 movq %rdx, %rcx 653 cmpq %r11, %rdx [all …]
|
A D | strcmp-evex.S | 146 cmpq %r11, %rdx 176 cmpq %r11, %rdx 212 cmpq %r11, %rdx 248 cmpq %r11, %rdx 322 subq %rdi, %rdx 331 subq %rdx, %r11 334 addq %rsi, %rdx 335 movq %rdx, %rsi 610 movq %rdx, %rcx 848 cmpq %r11, %rdx [all …]
|
A D | memcmpeq-evex.S | 103 cmpq $(VEC_SIZE * 2), %rdx 114 cmpq $(VEC_SIZE * 4), %rdx 131 cmpq $(VEC_SIZE * 8), %rdx 139 addq %rdx, %rdi 168 leaq -(VEC_SIZE * 4)(%rdi, %rdx), %rdx 196 cmpq %rdx, %rdi 199 subq %rdx, %rdi 213 VMOVU VEC_SIZE(%rsi, %rdx), %YMM2 214 vpxorq VEC_SIZE(%rdx), %YMM2, %YMM2 216 VMOVU (%rsi, %rdx), %YMM1 [all …]
|
A D | strcmp-sse2-unaligned.S | 41 bsfq %rax, %rdx 64 salq $16, %rdx 68 orq %rdx, %rax 70 movq %rcx, %rdx 71 salq $48, %rdx 72 orq %rdx, %rax 78 andq $-64, %rdx 79 subq %rdi, %rdx 92 addq $64, %rdx 148 movq %rdx, %r9 [all …]
|
A D | memcmpeq-avx2.S | 70 cmpq $(VEC_SIZE * 2), %rdx 83 cmpq $(VEC_SIZE * 4), %rdx 100 cmpq $(VEC_SIZE * 8), %rdx 109 addq %rdx, %rsi 112 addq %rdx, %rdi 197 leaq -(VEC_SIZE * 4)(%rdi, %rdx), %rdx 228 cmpq %rdx, %rdi 233 subq %rdx, %rdi 272 movl -4(%rdi, %rdx), %ecx 273 movl -4(%rsi, %rdx), %edi [all …]
|
A D | memcpy-ssse3.S | 195 lea -128(%rdx), %rdx 222 lea 0x80(%rdx), %rdx 271 lea 0x80(%rdx), %rdx 447 lea -64(%rdx), %rdx 475 lea 64(%rdx), %rdx 520 lea 64(%rdx), %rdx 560 lea 64(%rdx), %rdx 605 lea 64(%rdx), %rdx 645 lea 64(%rdx), %rdx 690 lea 64(%rdx), %rdx [all …]
|
A D | memcmp-avx2-movbe.S | 162 leal -1(%rdx, %rdx), %eax 179 leal -1(%rdx, %rdx), %eax 195 leal -1(%rdx, %rdx), %eax 227 leal -1(%rdx, %rdx), %eax 238 leaq -(VEC_SIZE * 4)(%rdi, %rdx), %rdx 270 cmpq %rdx, %rdi 273 subq %rdx, %rdi 353 leal -1(%rdx, %rdx), %eax 370 leal -1(%rdx, %rdx), %eax 387 leal -1(%rdx, %rdx), %eax [all …]
|
A D | strlen-vec.S | 58 orq %rcx, %rdx; 115 test %rdx, %rdx; \ 144 salq $16, %rdx 148 orq %rcx, %rdx 164 bts %rsi, %rdx 165 sarq %cl, %rdx 166 test %rdx, %rdx 203 bts %r11, %rdx 204 bsfq %rdx, %rdx 215 bsfq %rdx, %rdx [all …]
|
A D | memcmp-evex-movbe.S | 135 leal -1(%rdx, %rdx), %eax 229 movq %rdx, %rdi 257 leal -1(%rdx, %rdx), %eax 276 leal -1(%rdx, %rdx), %eax 294 leal -1(%rdx, %rdx), %eax 305 leaq -(VEC_SIZE * 4)(%rdi, %rdx, CHAR_SIZE), %rdx 330 cmpq %rdx, %rdi 333 subq %rdx, %rdi 390 leal -1(%rdx, %rdx), %eax 431 leal -1(%rdx, %rdx), %eax [all …]
|
A D | strcpy-sse2-unaligned.S | 78 test %rdx, %rdx 89 test %rdx, %rdx 115 test %rdx, %rdx 131 test %rdx, %rdx 147 test %rdx, %rdx 163 test %rdx, %rdx 179 test %rdx, %rdx 195 test %rdx, %rdx 227 test %rdx, %rdx 341 bsf %rdx, %rdx [all …]
|
A D | memcmp-sse4.S | 180 leal -1(%rdx, %rdx), %eax 199 leal -1(%rdx, %rdx), %eax 218 leal -1(%rdx, %rdx), %eax 302 add %rcx, %rdx 335 cmp $96, %rdx 431 cmp $64, %rdx 434 cmp $32, %rdx 464 sub $64, %rdx 487 sub $64, %rdx 545 sub $64, %rdx [all …]
|
A D | memmove-vec-unaligned-erms.S | 365 subq $4, %rdx 474 cmpq %rdx, %rcx 492 leaq (%rcx, %rdx), %r8 529 leaq (VEC_SIZE * -4)(%rcx, %rdx), %rdx 545 cmpq %rdi, %rdx 551 VMOVU %VEC(8), (%rdx) 658 cmpq %rdx, %rcx 735 cmpq %rcx, %rdx 768 addq %r8, %rdx 780 movq %rdx, %r10 [all …]
|
/sysdeps/x86_64/fpu/multiarch/ |
A D | svml_s_powf8_core_avx2.S | 120 vpand _ABSMASK(%rdx), %ymm9, %ymm8 122 vorpd _Two10(%rdx), %ymm3, %ymm2 123 vorpd _Two10(%rdx), %ymm13, %ymm3 138 vmovupd _One(%rdx), %ymm4 146 vandpd _Bias(%rdx), %ymm12, %ymm12 155 vandpd _Bias(%rdx), %ymm14, %ymm14 159 vorpd _Bias1(%rdx), %ymm12, %ymm11 197 vpcmpgtd _INF(%rdx), %ymm8, %ymm1 199 vpcmpeqd _INF(%rdx), %ymm8, %ymm8 212 vmulpd __dbC1(%rdx), %ymm7, %ymm4 [all …]
|
A D | svml_s_powf4_core_sse4.S | 113 orps _Two10(%rdx), %xmm6 115 orps _Two10(%rdx), %xmm10 144 subpd _One(%rdx), %xmm10 145 subpd _One(%rdx), %xmm6 163 andps _Bias(%rdx), %xmm9 165 andps _Bias(%rdx), %xmm8 167 orps _Bias1(%rdx), %xmm9 168 orps _Bias1(%rdx), %xmm8 173 mulpd _L2(%rdx), %xmm2 174 mulpd _L2(%rdx), %xmm1 [all …]
|
A D | svml_d_pow2_core_sse4.S | 131 orps _dbOne(%rdx), %xmm10 169 movups _LHN(%rdx), %xmm13 212 addpd _clv_3(%rdx), %xmm11 214 addpd _clv_4(%rdx), %xmm11 216 addpd _clv_5(%rdx), %xmm11 218 addpd _clv_6(%rdx), %xmm11 275 movq _INF(%rdx), %xmm7 302 movq _iOne(%rdx), %xmm3 320 addpd _cev_2(%rdx), %xmm3 322 addpd _cev_3(%rdx), %xmm3 [all …]
|