/sysdeps/x86_64/ |
A D | wcslen.S | 50 lea 16(%rax), %rax 57 lea 16(%rax), %rax 64 lea 16(%rax), %rax 70 lea 16(%rax), %rax 76 lea 16(%rax), %rax 82 lea 16(%rax), %rax 88 lea 16(%rax), %rax 94 lea 16(%rax), %rax 100 lea 16(%rax), %rax 106 lea 16(%rax), %rax [all …]
|
A D | strcat.S | 38 0: cmpb $0x0,(%rax) /* is byte NUL? */ 40 incq %rax /* increment pointer */ 110 incq %rax /* increment pointer */ 114 incq %rax /* increment pointer */ 118 incq %rax /* increment pointer */ 122 incq %rax /* increment pointer */ 128 incq %rax /* increment pointer */ 132 incq %rax /* increment pointer */ 136 incq %rax /* increment pointer */ 143 movq %rax, %rdx /* move around */ [all …]
|
A D | wcschr.S | 40 or %rax, %rdx 51 or %rax, %rdx 66 sar %cl, %rax 67 test %rax, %rax 70 bsf %rax, %rax 99 or %rax, %rdx 108 or %rax, %rdx 117 or %rax, %rdx 133 test %rax, %rax 135 bsf %rax, %rax [all …]
|
A D | strcspn.S | 44 movq %rsi, %rax /* Setup skipset. */ 56 movb 1(%rax), %cl /* get byte from skipset */ 61 movb 2(%rax), %cl /* get byte from skipset */ 72 L(1): leaq -4(%rdx), %rax /* prepare loop */ 86 movb (%rax), %cl /* get byte from string */ 90 movb 1(%rax), %cl /* get byte from string */ 94 movb 2(%rax), %cl /* get byte from string */ 102 incq %rax /* adjust pointer */ 103 L(6): incq %rax 104 L(5): incq %rax [all …]
|
A D | strspn.S | 43 movq %rsi, %rax /* Setup stopset. */ 55 movb 1(%rax), %cl /* get byte from stopset */ 60 movb 2(%rax), %cl /* get byte from stopset */ 71 L(1): leaq -4(%rdx), %rax /* prepare loop */ 85 movb (%rax), %cl /* get byte from string */ 89 movb 1(%rax), %cl /* get byte from string */ 93 movb 2(%rax), %cl /* get byte from string */ 97 movb 3(%rax), %cl /* get byte from string */ 101 incq %rax /* adjust pointer */ 102 L(6): incq %rax [all …]
|
A D | strcpy.S | 56 movq (%rsi), %rax /* Read double word (8 bytes). */ 58 movq %rax, %r9 /* Save a copy for NUL finding. */ 63 xorq %rax, %r9 /* (word+magic)^word */ 76 movq %rax, %r9 /* Save a copy for NUL finding. */ 81 xorq %rax, %r9 /* (word+magic)^word */ 94 movq %rax, %r9 /* Save a copy for NUL finding. */ 99 xorq %rax, %r9 /* (word+magic)^word */ 112 movq %rax, %r9 /* Save a copy for NUL finding. */ 117 xorq %rax, %r9 /* (word+magic)^word */ 142 shrq $16, %rax /* Shift... */ [all …]
|
A D | wcsrchr.S | 40 test %rax, %rax 55 mov %rax, %r8 71 shr %cl, %rax 74 test %rax, %rax 88 mov %rax, %r8 100 or %rax, %rcx 109 or %rax, %rcx 118 or %rax, %rcx 127 or %rax, %rcx 132 test %rax, %rax [all …]
|
A D | dl-tlsdesc.S | 41 movq 8(%rax), %rax 63 movq 8(%rax), %rax 64 subq %fs:0, %rax 107 movq TLSDESC_ARG(%rax), %rdi 108 movq (%rsi), %rax 111 movq TLSDESC_MODID(%rdi), %rax 112 salq $4, %rax 113 movq (%rax,%rsi), %rax 114 cmpq $-1, %rax 116 addq TLSDESC_MODOFF(%rdi), %rax [all …]
|
A D | strchr.S | 43 leaq (%rdi,%rax), %rax 46 leaq (%rdi,%rax), %rax 69 salq $32, %rax 71 orq %rcx, %rax 75 orq %rcx, %rax 76 testq %rax, %rax 123 orq %r8, %rax 129 bsfq %rax, %rax 131 leaq (%rdi,%rax), %rax 134 leaq (%rdi,%rax), %rax [all …]
|
A D | addmul_1.S | 40 mov (up), %rax 50 lea (%rax), %r11 57 lea (%rax), %r8 58 mov (up,%rbx,8), %rax 65 lea (%rax), %r8 69 lea (%rax), %r11 77 lea (%rax), %r8 86 lea (%rax), %r11 103 adc %rbp, %rax 107 ADDSUB %rax, %r10 [all …]
|
A D | mul_1.S | 41 add %r10, %rax 52 mov %rax, (rp) 59 mov %rax, %r9 68 mov %rax, %r8 75 mov %rax, %rbx 84 mov %rax, %r10 91 add %rax, %r9 97 add %rax, %r8 102 add %rax, %rbx 109 add %rax, %r10 [all …]
|
A D | strrchr.S | 25 movq %rdi, %rax 28 cmpq $4032, %rax 42 xorq %rdx, %rax 43 andq %rcx, %rax 45 bsrq %rax, %rax 46 addq %rdi, %rax 67 salq $32, %rax 69 orq %rdx, %rax 79 orq %rdx, %rax 149 orq %r9, %rax [all …]
|
A D | __longjmp.S | 50 shlq $32, %rax 51 orq %rax, %r9 64 rdsspq %rax 66 subq SHADOW_STACK_POINTER_OFFSET(%rdi), %rax 72 negq %rax 73 shrq $3, %rax 77 addq $1, %rax 81 cmpq %rbx, %rax 82 cmovb %rax, %rbx 84 subq %rbx, %rax
|
/sysdeps/x86_64/multiarch/ |
A D | strcat-sse2-unaligned.S | 39 xor %rax, %rax 50 mov %rdi, %rax 51 and $-16, %rax 54 mov %rdi, %rax 55 and $-16, %rax 58 sub %rax, %rcx 90 add $64, %rax 111 add $64, %rax 132 add $64, %rax 156 add $80, %rax [all …]
|
A D | strcat-avx2.S | 53 mov %rdi, %rax 57 mov %rdi, %rax 61 sub %rax, %rcx 220 sub %rdi, %rax 222 add %rdx, %rax 228 sub %rdi, %rax 231 add %rdx, %rax 236 sub %rdi, %rax 238 add %rdx, %rax 244 sub %rdi, %rax [all …]
|
A D | strcat-evex.S | 58 mov %rdi, %rax 62 mov %rdi, %rax 66 sub %rax, %rcx 225 sub %rdi, %rax 227 add %rdx, %rax 233 sub %rdi, %rax 236 add %rdx, %rax 241 sub %rdi, %rax 243 add %rdx, %rax 249 sub %rdi, %rax [all …]
|
A D | wcscpy-ssse3.S | 51 test %rax, %rax 81 test %rax, %rax 90 test %rax, %rax 99 test %rax, %rax 108 test %rax, %rax 117 test %rax, %rax 126 test %rax, %rax 153 test %rax, %rax 164 test %rax, %rax 171 test %rax, %rax [all …]
|
A D | strlen-vec.S | 64 xor %rax, %rax 89 movq %rdi, %rax 122 andq $-16, %rax 137 andq $-16, %rax 157 andq $-64, %rax 168 bsfq %rdx, %rax 181 addq $64, %rax 182 cmpq %rax, %r10 197 cmp %rax, %r11 205 addq %rdx, %rax [all …]
|
A D | strcpy-ssse3.S | 111 test %rax, %rax 181 test %rax, %rax 193 test %rax, %rax 205 test %rax, %rax 217 test %rax, %rax 229 test %rax, %rax 241 test %rax, %rax 274 test %rax, %rax 288 test %rax, %rax 297 test %rax, %rax [all …]
|
A D | strstr-sse2-unaligned.S | 30 movq %rdi, %rax 33 cmpq $4031, %rax 59 salq $16, %rax 60 orq %rax, %r8 63 bsf %r8, %rax 98 andq %rax, %r8 121 salq $32, %rax 124 orq %rax, %r8 127 bsfq %r8, %rax 227 orq %rax, %r8 [all …]
|
A D | strcat-ssse3.S | 84 lea 16(%rax), %rax 91 lea 16(%rax), %rax 98 lea 16(%rax), %rax 104 lea 16(%rax), %rax 110 lea 16(%rax), %rax 116 lea 16(%rax), %rax 122 lea 16(%rax), %rax 128 lea 16(%rax), %rax 134 lea 16(%rax), %rax 140 lea 16(%rax), %rax [all …]
|
A D | memchr-evex.S | 123 leaq (%rdi, %rax, CHAR_SIZE), %rax 140 leaq (%rdi, %rax, CHAR_SIZE), %rax 184 leaq (%RAW_PTR_REG, %rax, CHAR_SIZE), %rax 193 leaq VEC_SIZE(%rdi, %rax, CHAR_SIZE), %rax 199 leaq (VEC_SIZE * 2)(%rdi, %rax, CHAR_SIZE), %rax 205 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax 211 leaq (VEC_SIZE * 4)(%rdi, %rax, CHAR_SIZE), %rax 399 leaq (VEC_SIZE * 2)(%rdi, %rax, CHAR_SIZE), %rax 413 leaq VEC_SIZE(%rdi, %rax, CHAR_SIZE), %rax 461 tzcntq %rax, %rax [all …]
|
/sysdeps/x86_64/fpu/multiarch/ |
A D | svml_d_pow4_core_avx2.S | 116 vorpd _dbOne(%rax), %ymm4, %ymm13 122 vpaddd _HIDELTA(%rax), %xmm5, %xmm5 125 vpand _ABSMASK(%rax), %xmm2, %xmm10 128 vmovups _LORANGE(%rax), %xmm7 149 vaddpd _LHN(%rax), %ymm6, %ymm13 186 vmovupd _clv_2(%rax), %ymm8 224 vpand _ABSMASK(%rax), %xmm13, %xmm3 226 vpcmpgtd _INF(%rax), %xmm10, %xmm7 227 vpcmpeqd _INF(%rax), %xmm10, %xmm10 257 vpaddd _iOne(%rax), %xmm2, %xmm13 [all …]
|
A D | svml_d_pow8_core_avx512.S | 103 vmovups _dbOne(%rax), %zmm6 151 vaddpd _LHN(%rax), %zmm6, %zmm4 182 vmovups _clv_2(%rax), %zmm5 271 vmovups _cev_1(%rax), %zmm1 430 vmovdqu _ABSMASK(%rax), %ymm4 431 vmovdqu _LORANGE(%rax), %ymm3 434 vmovups _dbOne(%rax), %zmm11 435 vmovdqu _INF(%rax), %ymm5 463 vmovups _iHighMask(%rax), %zmm1 515 vmovups _clv_2(%rax), %zmm14 [all …]
|
A D | svml_s_sincosf8_core_avx2.S | 281 movl %r11d, (%rax) 332 vmovq %xmm0, %rax 336 shrq $32, %rax 342 shrq $32, %rax 348 shrq $32, %rax 354 shrq $32, %rax 357 vmovq %xmm0, %rax 361 shrq $32, %rax 367 shrq $32, %rax 373 shrq $32, %rax [all …]
|