Lines Matching refs:rcx

33 	lea	-16(%rcx, %r11), %r9;			\
49 lea -16(%rcx, %r11), %r9; \
123 mov LOCALE_T___LOCALES+LC_CTYPE*LP_SIZE(%rcx), %RAX_LP
125 mov (%rcx), %RAX_LP
144 and $0x3f, %rcx /* rsi alignment in cache line */
229 sub %rcx, %r9
264 mov $16, %rcx
274 movdqa (%rsi, %rcx), %xmm1
275 movdqa (%rdi, %rcx), %xmm2
289 add $16, %rcx
290 movdqa (%rsi, %rcx), %xmm1
291 movdqa (%rdi, %rcx), %xmm2
304 add $16, %rcx
331 mov $16, %rcx /* index for loads*/
348 movdqa (%rsi, %rcx), %xmm1
349 movdqa (%rdi, %rcx), %xmm2
372 add $16, %rcx
378 movdqa (%rsi, %rcx), %xmm1
379 movdqa (%rdi, %rcx), %xmm2
402 add $16, %rcx
432 movdqa (%rsi, %rcx), %xmm1
461 mov $16, %rcx /* index for loads */
478 movdqa (%rsi, %rcx), %xmm1
479 movdqa (%rdi, %rcx), %xmm2
503 add $16, %rcx
509 movdqa (%rsi, %rcx), %xmm1
510 movdqa (%rdi, %rcx), %xmm2
534 add $16, %rcx
556 movdqa (%rsi, %rcx), %xmm1
586 mov $16, %rcx /* index for loads */
603 movdqa (%rsi, %rcx), %xmm1
604 movdqa (%rdi, %rcx), %xmm2
628 add $16, %rcx
634 movdqa (%rsi, %rcx), %xmm1
635 movdqa (%rdi, %rcx), %xmm2
659 add $16, %rcx
681 movdqa (%rsi, %rcx), %xmm1
711 mov $16, %rcx /* index for loads */
728 movdqa (%rsi, %rcx), %xmm1
729 movdqa (%rdi, %rcx), %xmm2
753 add $16, %rcx
759 movdqa (%rsi, %rcx), %xmm1
760 movdqa (%rdi, %rcx), %xmm2
784 add $16, %rcx
806 movdqa (%rsi, %rcx), %xmm1
836 mov $16, %rcx /* index for loads */
853 movdqa (%rsi, %rcx), %xmm1
854 movdqa (%rdi, %rcx), %xmm2
878 add $16, %rcx
884 movdqa (%rsi, %rcx), %xmm1
885 movdqa (%rdi, %rcx), %xmm2
909 add $16, %rcx
931 movdqa (%rsi, %rcx), %xmm1
961 mov $16, %rcx /* index for loads */
978 movdqa (%rsi, %rcx), %xmm1
979 movdqa (%rdi, %rcx), %xmm2
1003 add $16, %rcx
1009 movdqa (%rsi, %rcx), %xmm1
1010 movdqa (%rdi, %rcx), %xmm2
1034 add $16, %rcx
1056 movdqa (%rsi, %rcx), %xmm1
1086 mov $16, %rcx /* index for loads */
1103 movdqa (%rsi, %rcx), %xmm1
1104 movdqa (%rdi, %rcx), %xmm2
1128 add $16, %rcx
1134 movdqa (%rsi, %rcx), %xmm1
1135 movdqa (%rdi, %rcx), %xmm2
1159 add $16, %rcx
1181 movdqa (%rsi, %rcx), %xmm1
1211 mov $16, %rcx /* index for loads */
1228 movdqa (%rsi, %rcx), %xmm1
1229 movdqa (%rdi, %rcx), %xmm2
1253 add $16, %rcx
1259 movdqa (%rsi, %rcx), %xmm1
1260 movdqa (%rdi, %rcx), %xmm2
1284 add $16, %rcx
1306 movdqa (%rsi, %rcx), %xmm1
1336 mov $16, %rcx /* index for loads */
1353 movdqa (%rsi, %rcx), %xmm1
1354 movdqa (%rdi, %rcx), %xmm2
1378 add $16, %rcx
1384 movdqa (%rsi, %rcx), %xmm1
1385 movdqa (%rdi, %rcx), %xmm2
1409 add $16, %rcx
1431 movdqa (%rsi, %rcx), %xmm1
1461 mov $16, %rcx /* index for loads */
1478 movdqa (%rsi, %rcx), %xmm1
1479 movdqa (%rdi, %rcx), %xmm2
1503 add $16, %rcx
1509 movdqa (%rsi, %rcx), %xmm1
1510 movdqa (%rdi, %rcx), %xmm2
1534 add $16, %rcx
1556 movdqa (%rsi, %rcx), %xmm1
1586 mov $16, %rcx /* index for loads */
1603 movdqa (%rsi, %rcx), %xmm1
1604 movdqa (%rdi, %rcx), %xmm2
1628 add $16, %rcx
1634 movdqa (%rsi, %rcx), %xmm1
1635 movdqa (%rdi, %rcx), %xmm2
1659 add $16, %rcx
1681 movdqa (%rsi, %rcx), %xmm1
1711 mov $16, %rcx /* index for loads */
1728 movdqa (%rsi, %rcx), %xmm1
1729 movdqa (%rdi, %rcx), %xmm2
1753 add $16, %rcx
1759 movdqa (%rsi, %rcx), %xmm1
1760 movdqa (%rdi, %rcx), %xmm2
1784 add $16, %rcx
1806 movdqa (%rsi, %rcx), %xmm1
1836 mov $16, %rcx /* index for loads */
1853 movdqa (%rsi, %rcx), %xmm1
1854 movdqa (%rdi, %rcx), %xmm2
1878 add $16, %rcx
1884 movdqa (%rsi, %rcx), %xmm1
1885 movdqa (%rdi, %rcx), %xmm2
1909 add $16, %rcx
1931 movdqa (%rsi, %rcx), %xmm1
1961 mov $16, %rcx /* index for loads */
1978 movdqa (%rsi, %rcx), %xmm1
1979 movdqa (%rdi, %rcx), %xmm2
2003 add $16, %rcx
2009 movdqa (%rsi, %rcx), %xmm1
2010 movdqa (%rdi, %rcx), %xmm2
2034 add $16, %rcx
2056 movdqa (%rsi, %rcx), %xmm1
2087 mov $16, %rcx /* index for loads */
2105 movdqa (%rsi, %rcx), %xmm1
2106 movdqa (%rdi, %rcx), %xmm2
2130 add $16, %rcx
2136 movdqa (%rsi, %rcx), %xmm1
2137 movdqa (%rdi, %rcx), %xmm2
2161 add $16, %rcx
2183 movdqa (%rsi, %rcx), %xmm1
2197 lea -16(%r9, %rcx), %rax /* locate the exact offset for rdi */
2200 lea (%rsi, %rcx), %rsi /* locate the exact address for second operand(rsi) */
2219 movl (%rdx,%rcx,4), %ecx
2237 movl (%rdx,%rcx,4), %ecx