/sysdeps/x86_64/fpu/ |
A D | svml_s_sincosf16_core.S | 159 movl %esp, %ebp 179 movl -240(%ebp), %eax 182 movl -236(%ebp), %eax 185 movl -232(%ebp), %eax 188 movl -228(%ebp), %eax 191 movl -224(%ebp), %eax 194 movl -220(%ebp), %eax 197 movl -216(%ebp), %eax 200 movl -212(%ebp), %eax 203 movl -208(%ebp), %eax [all …]
|
A D | svml_d_sincos8_core.S | 109 movl %esp, %ebp 129 movl -208(%ebp), %eax 132 movl -204(%ebp), %eax 135 movl -200(%ebp), %eax 138 movl -196(%ebp), %eax 141 movl -192(%ebp), %eax 144 movl -188(%ebp), %eax 147 movl -184(%ebp), %eax 150 movl -180(%ebp), %eax 153 movl -240(%ebp), %eax [all …]
|
A D | svml_s_sincosf8_core.S | 113 movl %esp, %ebp 134 movl -144(%ebp), %eax 137 movl -140(%ebp), %eax 140 movl -136(%ebp), %eax 143 movl -132(%ebp), %eax 146 movl -128(%ebp), %eax 147 vmovss -96(%ebp), %xmm0 149 movl -124(%ebp), %eax 150 vmovss -92(%ebp), %xmm0 152 movl -120(%ebp), %eax [all …]
|
A D | svml_s_sincosf8_core_avx.S | 109 movl %esp, %ebp 132 movq -128(%ebp), %rax 140 movq -120(%ebp), %rax 147 movq -144(%ebp), %rax 148 vmovss -96(%ebp), %xmm0 150 vmovss -92(%ebp), %xmm0 153 movq -136(%ebp), %rax 154 vmovss -88(%ebp), %xmm0 156 vmovss -84(%ebp), %xmm0 159 movq -160(%ebp), %rax [all …]
|
A D | svml_d_sincos4_core.S | 87 movl %esp, %ebp 108 movq -128(%ebp), %rax 116 movq -120(%ebp), %rax 117 vmovsd -96(%ebp), %xmm0 119 vmovsd -88(%ebp), %xmm0 122 movq -144(%ebp), %rax 123 vmovsd -80(%ebp), %xmm0 125 vmovsd -72(%ebp), %xmm0 128 movq -136(%ebp), %rax 129 vmovsd -64(%ebp), %xmm0 [all …]
|
A D | svml_d_sincos4_core_avx.S | 82 movl %esp, %ebp 103 movq -128(%ebp), %rax 111 movq -120(%ebp), %rax 112 vmovsd -96(%ebp), %xmm0 114 vmovsd -88(%ebp), %xmm0 117 movq -144(%ebp), %rax 118 vmovsd -80(%ebp), %xmm0 120 vmovsd -72(%ebp), %xmm0 123 movq -136(%ebp), %rax 124 vmovsd -64(%ebp), %xmm0 [all …]
|
/sysdeps/i386/i586/ |
A D | addmul_1.S | 41 pushl %ebp 56 xorl %ebp, %ebp 57 cfi_rel_offset (ebp, 4) 60 L(oop): adcl $0, %ebp 65 addl %ebp, %eax 69 addl %eax, %ebp 74 movl %edx, %ebp 77 adcl $0, %ebp 78 movl %ebp, %eax 82 popl %ebp [all …]
|
A D | submul_1.S | 41 pushl %ebp 56 xorl %ebp, %ebp 57 cfi_rel_offset (ebp, 4) 60 L(oop): adcl $0, %ebp 65 addl %ebp, %eax 69 subl %eax, %ebp 74 movl %edx, %ebp 77 adcl $0, %ebp 78 movl %ebp, %eax 82 popl %ebp [all …]
|
A D | add_n.S | 36 pushl %ebp 66 adcl %ebp,%eax 68 adcl %ebp,%edx 75 adcl %ebp,%eax 77 adcl %ebp,%edx 84 adcl %ebp,%eax 86 adcl %ebp,%edx 93 adcl %ebp,%eax 95 adcl %ebp,%edx 114 adcl %ebp,%eax [all …]
|
A D | sub_n.S | 36 pushl %ebp 66 sbbl %ebp,%eax 68 sbbl %ebp,%edx 75 sbbl %ebp,%eax 77 sbbl %ebp,%edx 84 sbbl %ebp,%eax 86 sbbl %ebp,%edx 93 sbbl %ebp,%eax 95 sbbl %ebp,%edx 114 sbbl %ebp,%eax [all …]
|
A D | mul_1.S | 41 pushl %ebp 56 xorl %ebp, %ebp 57 cfi_rel_offset (ebp, 4) 60 L(oop): adcl $0, %ebp 65 addl %eax, %ebp 67 movl %ebp, (%res_ptr,%size,4) 70 movl %edx, %ebp 73 adcl $0, %ebp 74 movl %ebp, %eax 78 popl %ebp [all …]
|
A D | lshift.S | 35 pushl %ebp 80 movl %edx,%ebp 86 movl %ebp,(%edi) 139 popl %ebp 180 movl %edx,%ebp 185 movl %ebp,(%edi) 191 adcl %ebp,%ebp 200 adcl %ebp,%ebp 223 movl %edx,%ebp 226 movl %ebp,(%edi) [all …]
|
A D | strchr.S | 51 pushl %ebp 118 movl $magic, %ebp 121 addl %ecx, %ebp 185 addl $1, %ebp 197 movl %edi, %ebp 200 addl %ecx, %ebp 206 xorl %ecx, %ebp 213 orl $magic, %ebp 215 addl $1, %ebp 257 movl %edi, %ebp [all …]
|
A D | rshift.S | 35 pushl %ebp 77 movl %edx,%ebp 83 movl %ebp,(%edi) 86 movl 8(%esi),%ebp 136 popl %ebp 180 movl %edx,%ebp 185 movl %ebp,(%edi) 191 rcrl $1,%ebp 200 rcrl $1,%ebp 223 movl %edx,%ebp [all …]
|
/sysdeps/i386/ |
A D | backtrace.c | 74 struct layout *ebp; member 100 struct layout *ebp = (struct layout *) arg.lastebp; in __backtrace() local 105 if ((void *) ebp < arg.lastesp || (void *) ebp > __libc_stack_end in __backtrace() 106 || ((long) ebp & 3)) in __backtrace() 109 array[arg.cnt++] = ebp->ret; in __backtrace() 110 ebp = ebp->ebp; in __backtrace()
|
A D | mul_1.S | 41 pushl %ebp 56 xorl %ebp, %ebp 57 cfi_rel_offset (ebp, 4) 62 addl %ebp, %eax 65 movl %edx, %ebp 69 movl %ebp, %eax 74 popl %ebp 76 cfi_restore (ebp)
|
A D | submul_1.S | 41 pushl %ebp 47 cfi_rel_offset (ebp, 4) 57 xorl %ebp, %ebp 62 addl %ebp, %eax 66 movl %edx, %ebp 70 movl %ebp, %eax 75 popl %ebp 77 cfi_restore (ebp)
|
A D | addmul_1.S | 41 pushl %ebp 56 xorl %ebp, %ebp 57 cfi_rel_offset (ebp, 4) 62 addl %ebp, %eax 66 movl %edx, %ebp 70 movl %ebp, %eax 75 popl %ebp 77 cfi_restore (ebp)
|
A D | sysdep.h | 45 pushl %ebp; cfi_adjust_cfa_offset (4); movl %esp, %ebp; \ 46 cfi_def_cfa_register (ebp); call JUMPTARGET(mcount); \ 47 popl %ebp; cfi_def_cfa (esp, 4);
|
/sysdeps/unix/sysv/linux/i386/ |
A D | libc-do-syscall.S | 35 pushl %ebp 37 cfi_rel_offset (ebp, 0) 39 movl 8(%edi), %ebp 42 popl %ebp 44 cfi_restore (ebp)
|
A D | configure.ac | 4 # Check if CFLAGS allows compiler to use ebp register in inline assembly. 5 AC_CACHE_CHECK([if compiler flags allows ebp in inline assembly], 11 register int reg asm ("ebp") = i;
|
A D | configure | 4 # Check if CFLAGS allows compiler to use ebp register in inline assembly. 6 { $as_echo "$as_me:${as_lineno-$LINENO}: checking if compiler flags allows ebp in inline assembly" … 7 $as_echo_n "checking if compiler flags allows ebp in inline assembly... " >&6; } 17 register int reg asm ("ebp") = i;
|
/sysdeps/x86_64/fpu/multiarch/ |
A D | svml_s_sincosf16_core_avx512.S | 621 movl %esp, %ebp 630 movl -240(%ebp), %eax 633 movl -236(%ebp), %eax 636 movl -232(%ebp), %eax 639 movl -228(%ebp), %eax 642 movl -224(%ebp), %eax 645 movl -220(%ebp), %eax 648 movl -216(%ebp), %eax 651 movl -212(%ebp), %eax 654 movl -208(%ebp), %eax [all …]
|
A D | svml_s_sincosf8_core_avx2.S | 322 movl %esp, %ebp 333 vmovss -80(%ebp), %xmm0 335 vmovss -76(%ebp), %xmm0 338 movq -104(%ebp), %rax 339 vmovss -72(%ebp), %xmm0 341 vmovss -68(%ebp), %xmm0 344 movq -96(%ebp), %rax 345 vmovss -64(%ebp), %xmm0 347 vmovss -60(%ebp), %xmm0 350 movq -88(%ebp), %rax [all …]
|
A D | svml_d_sincos4_core_avx2.S | 332 movl %esp, %ebp 338 vmovaps %xmm1, -96(%ebp) 341 movl -96(%ebp), %eax 342 vmovsd -80(%ebp), %xmm0 344 movl -92(%ebp), %eax 345 vmovsd -72(%ebp), %xmm0 347 movl -88(%ebp), %eax 348 vmovsd -64(%ebp), %xmm0 350 movl -84(%ebp), %eax 351 vmovsd -56(%ebp), %xmm0 [all …]
|