2 /* Do not modify. This file is auto-generated from bsaes-x86_64.pl. */
8 .type _bsaes_encrypt8,@function
20 .byte 102,68,15,56,0,255
32 _bsaes_encrypt8_bitslice:
128 .byte 102,68,15,56,0,255
129 .byte 102,15,56,0,199
132 .byte 102,15,56,0,207
133 .byte 102,15,56,0,215
136 .byte 102,15,56,0,223
137 .byte 102,15,56,0,231
138 .byte 102,15,56,0,239
139 .byte 102,15,56,0,247
329 pshufd $0x93,%xmm15,%xmm7
330 pshufd $0x93,%xmm0,%xmm8
332 pshufd $0x93,%xmm3,%xmm9
334 pshufd $0x93,%xmm5,%xmm10
336 pshufd $0x93,%xmm2,%xmm11
338 pshufd $0x93,%xmm6,%xmm12
340 pshufd $0x93,%xmm1,%xmm13
342 pshufd $0x93,%xmm4,%xmm14
349 pshufd $0x4E,%xmm15,%xmm15
351 pshufd $0x4E,%xmm0,%xmm0
357 pshufd $0x4E,%xmm2,%xmm7
359 pshufd $0x4E,%xmm6,%xmm8
361 pshufd $0x4E,%xmm5,%xmm2
363 pshufd $0x4E,%xmm4,%xmm6
365 pshufd $0x4E,%xmm1,%xmm5
367 pshufd $0x4E,%xmm3,%xmm1
375 movdqa 48(%r11),%xmm7
377 movdqa 64(%r11),%xmm7
382 movdqa 16(%r11),%xmm8
411 movdqa 32(%r11),%xmm7
478 .size _bsaes_encrypt8,.-_bsaes_encrypt8
480 .type _bsaes_decrypt8,@function
483 leaq .LBS0(%rip),%r11
487 movdqa -48(%r11),%xmm7
492 .byte 102,68,15,56,0,255
493 .byte 102,15,56,0,199
496 .byte 102,15,56,0,207
497 .byte 102,15,56,0,215
500 .byte 102,15,56,0,223
501 .byte 102,15,56,0,231
502 .byte 102,15,56,0,239
503 .byte 102,15,56,0,247
505 movdqa 16(%r11),%xmm8
534 movdqa 32(%r11),%xmm7
599 .byte 102,68,15,56,0,255
600 .byte 102,15,56,0,199
603 .byte 102,15,56,0,207
604 .byte 102,15,56,0,215
607 .byte 102,15,56,0,223
608 .byte 102,15,56,0,231
609 .byte 102,15,56,0,239
610 .byte 102,15,56,0,247
801 pshufd $0x4E,%xmm15,%xmm7
802 pshufd $0x4E,%xmm2,%xmm13
804 pshufd $0x4E,%xmm4,%xmm14
806 pshufd $0x4E,%xmm0,%xmm8
808 pshufd $0x4E,%xmm5,%xmm9
810 pshufd $0x4E,%xmm3,%xmm10
814 pshufd $0x4E,%xmm1,%xmm11
818 pshufd $0x4E,%xmm6,%xmm12
832 pshufd $0x93,%xmm15,%xmm7
833 pshufd $0x93,%xmm0,%xmm8
835 pshufd $0x93,%xmm5,%xmm9
837 pshufd $0x93,%xmm3,%xmm10
839 pshufd $0x93,%xmm1,%xmm11
841 pshufd $0x93,%xmm6,%xmm12
843 pshufd $0x93,%xmm2,%xmm13
845 pshufd $0x93,%xmm4,%xmm14
852 pshufd $0x4E,%xmm15,%xmm15
854 pshufd $0x4E,%xmm0,%xmm0
860 pshufd $0x4E,%xmm1,%xmm7
862 pshufd $0x4E,%xmm6,%xmm8
864 pshufd $0x4E,%xmm3,%xmm1
866 pshufd $0x4E,%xmm4,%xmm6
868 pshufd $0x4E,%xmm2,%xmm3
870 pshufd $0x4E,%xmm5,%xmm2
881 movdqa -16(%r11),%xmm7
883 movdqa -32(%r11),%xmm7
888 movdqa 16(%r11),%xmm8
917 movdqa 32(%r11),%xmm7
984 .size _bsaes_decrypt8,.-_bsaes_decrypt8
985 .type _bsaes_key_convert,@function
988 leaq .Lmasks(%rip),%r11
992 movdqa 16(%r11),%xmm1
993 movdqa 32(%r11),%xmm2
994 movdqa 48(%r11),%xmm3
995 movdqa 64(%r11),%xmm4
1005 .byte 102,15,56,0,244
1022 pcmpeqb %xmm2,%xmm10
1025 pcmpeqb %xmm3,%xmm11
1035 movdqa %xmm8,0(%rax)
1036 pcmpeqb %xmm0,%xmm12
1038 movdqa %xmm9,16(%rax)
1039 pcmpeqb %xmm1,%xmm13
1045 movdqa %xmm10,32(%rax)
1046 pcmpeqb %xmm2,%xmm14
1048 movdqa %xmm11,48(%rax)
1049 pcmpeqb %xmm3,%xmm15
1055 movdqa %xmm12,64(%rax)
1056 movdqa %xmm13,80(%rax)
1057 movdqa %xmm14,96(%rax)
1058 movdqa %xmm15,112(%rax)
1063 movdqa 80(%r11),%xmm7
1066 .size _bsaes_key_convert,.-_bsaes_key_convert
1068 .globl bsaes_cbc_encrypt
1069 .type bsaes_cbc_encrypt,@function
1074 jne asm_AES_cbc_encrypt
1076 jb asm_AES_cbc_encrypt
1081 .cfi_adjust_cfa_offset 8
1082 .cfi_offset %rbp,-16
1084 .cfi_adjust_cfa_offset 8
1085 .cfi_offset %rbx,-24
1087 .cfi_adjust_cfa_offset 8
1088 .cfi_offset %r12,-32
1090 .cfi_adjust_cfa_offset 8
1091 .cfi_offset %r13,-40
1093 .cfi_adjust_cfa_offset 8
1094 .cfi_offset %r14,-48
1096 .cfi_adjust_cfa_offset 8
1097 .cfi_offset %r15,-56
1099 .cfi_adjust_cfa_offset 0x48
1101 .cfi_def_cfa_register %rbp
1118 call _bsaes_key_convert
1123 movdqu (%rbx),%xmm14
1126 movdqu 0(%r12),%xmm15
1127 movdqu 16(%r12),%xmm0
1128 movdqu 32(%r12),%xmm1
1129 movdqu 48(%r12),%xmm2
1130 movdqu 64(%r12),%xmm3
1131 movdqu 80(%r12),%xmm4
1133 movdqu 96(%r12),%xmm5
1135 movdqu 112(%r12),%xmm6
1136 movdqa %xmm14,32(%rbp)
1138 call _bsaes_decrypt8
1140 pxor 32(%rbp),%xmm15
1141 movdqu 0(%r12),%xmm7
1142 movdqu 16(%r12),%xmm8
1144 movdqu 32(%r12),%xmm9
1146 movdqu 48(%r12),%xmm10
1148 movdqu 64(%r12),%xmm11
1150 movdqu 80(%r12),%xmm12
1152 movdqu 96(%r12),%xmm13
1154 movdqu 112(%r12),%xmm14
1156 movdqu %xmm15,0(%r13)
1158 movdqu %xmm0,16(%r13)
1159 movdqu %xmm5,32(%r13)
1160 movdqu %xmm3,48(%r13)
1161 movdqu %xmm1,64(%r13)
1162 movdqu %xmm6,80(%r13)
1163 movdqu %xmm2,96(%r13)
1164 movdqu %xmm4,112(%r13)
1172 movdqu 0(%r12),%xmm15
1177 movdqu 16(%r12),%xmm0
1179 movdqu 32(%r12),%xmm1
1182 movdqu 48(%r12),%xmm2
1184 movdqu 64(%r12),%xmm3
1187 movdqu 80(%r12),%xmm4
1189 movdqu 96(%r12),%xmm5
1190 movdqa %xmm14,32(%rbp)
1191 call _bsaes_decrypt8
1192 pxor 32(%rbp),%xmm15
1193 movdqu 0(%r12),%xmm7
1194 movdqu 16(%r12),%xmm8
1196 movdqu 32(%r12),%xmm9
1198 movdqu 48(%r12),%xmm10
1200 movdqu 64(%r12),%xmm11
1202 movdqu 80(%r12),%xmm12
1204 movdqu 96(%r12),%xmm14
1206 movdqu %xmm15,0(%r13)
1207 movdqu %xmm0,16(%r13)
1208 movdqu %xmm5,32(%r13)
1209 movdqu %xmm3,48(%r13)
1210 movdqu %xmm1,64(%r13)
1211 movdqu %xmm6,80(%r13)
1212 movdqu %xmm2,96(%r13)
1216 movdqa %xmm14,32(%rbp)
1217 call _bsaes_decrypt8
1218 pxor 32(%rbp),%xmm15
1219 movdqu 0(%r12),%xmm7
1220 movdqu 16(%r12),%xmm8
1222 movdqu 32(%r12),%xmm9
1224 movdqu 48(%r12),%xmm10
1226 movdqu 64(%r12),%xmm11
1228 movdqu 80(%r12),%xmm14
1230 movdqu %xmm15,0(%r13)
1231 movdqu %xmm0,16(%r13)
1232 movdqu %xmm5,32(%r13)
1233 movdqu %xmm3,48(%r13)
1234 movdqu %xmm1,64(%r13)
1235 movdqu %xmm6,80(%r13)
1239 movdqa %xmm14,32(%rbp)
1240 call _bsaes_decrypt8
1241 pxor 32(%rbp),%xmm15
1242 movdqu 0(%r12),%xmm7
1243 movdqu 16(%r12),%xmm8
1245 movdqu 32(%r12),%xmm9
1247 movdqu 48(%r12),%xmm10
1249 movdqu 64(%r12),%xmm14
1251 movdqu %xmm15,0(%r13)
1252 movdqu %xmm0,16(%r13)
1253 movdqu %xmm5,32(%r13)
1254 movdqu %xmm3,48(%r13)
1255 movdqu %xmm1,64(%r13)
1259 movdqa %xmm14,32(%rbp)
1260 call _bsaes_decrypt8
1261 pxor 32(%rbp),%xmm15
1262 movdqu 0(%r12),%xmm7
1263 movdqu 16(%r12),%xmm8
1265 movdqu 32(%r12),%xmm9
1267 movdqu 48(%r12),%xmm14
1269 movdqu %xmm15,0(%r13)
1270 movdqu %xmm0,16(%r13)
1271 movdqu %xmm5,32(%r13)
1272 movdqu %xmm3,48(%r13)
1276 movdqa %xmm14,32(%rbp)
1277 call _bsaes_decrypt8
1278 pxor 32(%rbp),%xmm15
1279 movdqu 0(%r12),%xmm7
1280 movdqu 16(%r12),%xmm8
1282 movdqu 32(%r12),%xmm14
1284 movdqu %xmm15,0(%r13)
1285 movdqu %xmm0,16(%r13)
1286 movdqu %xmm5,32(%r13)
1290 movdqa %xmm14,32(%rbp)
1291 call _bsaes_decrypt8
1292 pxor 32(%rbp),%xmm15
1293 movdqu 0(%r12),%xmm7
1294 movdqu 16(%r12),%xmm14
1296 movdqu %xmm15,0(%r13)
1297 movdqu %xmm0,16(%r13)
1304 call asm_AES_decrypt
1305 pxor 32(%rbp),%xmm14
1306 movdqu %xmm14,(%r13)
1307 movdqa %xmm15,%xmm14
1310 movdqu %xmm14,(%rbx)
1314 movdqa %xmm0,0(%rax)
1315 movdqa %xmm0,16(%rax)
1335 .cfi_def_cfa_register %rsp
1339 .size bsaes_cbc_encrypt,.-bsaes_cbc_encrypt
1341 .globl bsaes_ctr32_encrypt_blocks
1342 .type bsaes_ctr32_encrypt_blocks,@function
1344 bsaes_ctr32_encrypt_blocks:
1349 .cfi_adjust_cfa_offset 8
1350 .cfi_offset %rbp,-16
1352 .cfi_adjust_cfa_offset 8
1353 .cfi_offset %rbx,-24
1355 .cfi_adjust_cfa_offset 8
1356 .cfi_offset %r12,-32
1358 .cfi_adjust_cfa_offset 8
1359 .cfi_offset %r13,-40
1361 .cfi_adjust_cfa_offset 8
1362 .cfi_offset %r14,-48
1364 .cfi_adjust_cfa_offset 8
1365 .cfi_offset %r15,-56
1367 .cfi_adjust_cfa_offset 0x48
1369 .cfi_def_cfa_register %rbp
1376 movdqa %xmm0,32(%rbp)
1388 call _bsaes_key_convert
1393 leaq .LADD1(%rip),%r11
1394 movdqa 32(%rbp),%xmm15
1395 movdqa -32(%r11),%xmm7
1396 .byte 102,68,15,56,0,199
1397 .byte 102,68,15,56,0,255
1402 movdqa %xmm15,32(%rbp)
1407 paddd 16(%r11),%xmm1
1409 paddd 32(%r11),%xmm2
1411 paddd 48(%r11),%xmm3
1413 paddd 64(%r11),%xmm4
1415 paddd 80(%r11),%xmm5
1416 paddd 96(%r11),%xmm6
1422 movdqa -16(%r11),%xmm7
1427 .byte 102,68,15,56,0,255
1428 .byte 102,15,56,0,199
1431 .byte 102,15,56,0,207
1432 .byte 102,15,56,0,215
1435 .byte 102,15,56,0,223
1436 .byte 102,15,56,0,231
1437 .byte 102,15,56,0,239
1438 .byte 102,15,56,0,247
1439 leaq .LBS0(%rip),%r11
1442 call _bsaes_encrypt8_bitslice
1445 jc .Lctr_enc_loop_done
1447 movdqu 0(%r12),%xmm7
1448 movdqu 16(%r12),%xmm8
1449 movdqu 32(%r12),%xmm9
1450 movdqu 48(%r12),%xmm10
1451 movdqu 64(%r12),%xmm11
1452 movdqu 80(%r12),%xmm12
1453 movdqu 96(%r12),%xmm13
1454 movdqu 112(%r12),%xmm14
1457 movdqa 32(%rbp),%xmm15
1459 movdqu %xmm7,0(%r13)
1461 movdqu %xmm0,16(%r13)
1463 movdqu %xmm3,32(%r13)
1465 movdqu %xmm5,48(%r13)
1467 movdqu %xmm2,64(%r13)
1469 movdqu %xmm6,80(%r13)
1471 movdqu %xmm1,96(%r13)
1472 leaq .LADD1(%rip),%r11
1473 movdqu %xmm4,112(%r13)
1475 paddd 112(%r11),%xmm15
1480 .Lctr_enc_loop_done:
1482 movdqu 0(%r12),%xmm7
1484 movdqu %xmm15,0(%r13)
1487 movdqu 16(%r12),%xmm8
1489 movdqu %xmm0,16(%r13)
1491 movdqu 32(%r12),%xmm9
1493 movdqu %xmm3,32(%r13)
1496 movdqu 48(%r12),%xmm10
1498 movdqu %xmm5,48(%r13)
1500 movdqu 64(%r12),%xmm11
1502 movdqu %xmm2,64(%r13)
1505 movdqu 80(%r12),%xmm12
1507 movdqu %xmm6,80(%r13)
1509 movdqu 96(%r12),%xmm13
1511 movdqu %xmm1,96(%r13)
1519 call asm_AES_encrypt
1537 movdqa %xmm0,0(%rax)
1538 movdqa %xmm0,16(%rax)
1558 .cfi_def_cfa_register %rsp
1562 .size bsaes_ctr32_encrypt_blocks,.-bsaes_ctr32_encrypt_blocks
1563 .globl bsaes_xts_encrypt
1564 .type bsaes_xts_encrypt,@function
1571 .cfi_adjust_cfa_offset 8
1572 .cfi_offset %rbp,-16
1574 .cfi_adjust_cfa_offset 8
1575 .cfi_offset %rbx,-24
1577 .cfi_adjust_cfa_offset 8
1578 .cfi_offset %r12,-32
1580 .cfi_adjust_cfa_offset 8
1581 .cfi_offset %r13,-40
1583 .cfi_adjust_cfa_offset 8
1584 .cfi_offset %r14,-48
1586 .cfi_adjust_cfa_offset 8
1587 .cfi_offset %r15,-56
1589 .cfi_adjust_cfa_offset 0x48
1591 .cfi_def_cfa_register %rbp
1600 call asm_AES_encrypt
1613 call _bsaes_key_convert
1619 movdqa 32(%rbp),%xmm6
1622 movdqa .Lxts_magic(%rip),%xmm12
1623 pcmpgtd %xmm6,%xmm14
1631 pshufd $0x13,%xmm14,%xmm13
1634 movdqa %xmm6,0(%rsp)
1637 pcmpgtd %xmm6,%xmm14
1639 pshufd $0x13,%xmm14,%xmm13
1642 movdqa %xmm6,16(%rsp)
1645 pcmpgtd %xmm6,%xmm14
1647 movdqu 0(%r12),%xmm7
1648 pshufd $0x13,%xmm14,%xmm13
1651 movdqa %xmm6,32(%rsp)
1654 pcmpgtd %xmm6,%xmm14
1656 movdqu 16(%r12),%xmm8
1658 pshufd $0x13,%xmm14,%xmm13
1661 movdqa %xmm6,48(%rsp)
1664 pcmpgtd %xmm6,%xmm14
1666 movdqu 32(%r12),%xmm9
1668 pshufd $0x13,%xmm14,%xmm13
1671 movdqa %xmm6,64(%rsp)
1674 pcmpgtd %xmm6,%xmm14
1676 movdqu 48(%r12),%xmm10
1678 pshufd $0x13,%xmm14,%xmm13
1681 movdqa %xmm6,80(%rsp)
1684 pcmpgtd %xmm6,%xmm14
1686 movdqu 64(%r12),%xmm11
1688 pshufd $0x13,%xmm14,%xmm13
1691 movdqa %xmm6,96(%rsp)
1694 pcmpgtd %xmm6,%xmm14
1696 movdqu 80(%r12),%xmm12
1698 movdqu 96(%r12),%xmm13
1700 movdqu 112(%r12),%xmm14
1702 movdqa %xmm6,112(%rsp)
1708 call _bsaes_encrypt8
1712 movdqu %xmm15,0(%r13)
1714 movdqu %xmm0,16(%r13)
1716 movdqu %xmm3,32(%r13)
1718 movdqu %xmm5,48(%r13)
1720 movdqu %xmm2,64(%r13)
1722 movdqu %xmm6,80(%r13)
1723 pxor 112(%rsp),%xmm4
1724 movdqu %xmm1,96(%r13)
1725 movdqu %xmm4,112(%r13)
1728 movdqa 112(%rsp),%xmm6
1730 movdqa .Lxts_magic(%rip),%xmm12
1731 pcmpgtd %xmm6,%xmm14
1732 pshufd $0x13,%xmm14,%xmm13
1736 pcmpgtd %xmm6,%xmm14
1745 pshufd $0x13,%xmm14,%xmm13
1748 movdqa %xmm6,0(%rsp)
1751 pcmpgtd %xmm6,%xmm14
1753 pshufd $0x13,%xmm14,%xmm13
1756 movdqa %xmm6,16(%rsp)
1759 pcmpgtd %xmm6,%xmm14
1761 movdqu 0(%r12),%xmm7
1764 pshufd $0x13,%xmm14,%xmm13
1767 movdqa %xmm6,32(%rsp)
1770 pcmpgtd %xmm6,%xmm14
1772 movdqu 16(%r12),%xmm8
1776 pshufd $0x13,%xmm14,%xmm13
1779 movdqa %xmm6,48(%rsp)
1782 pcmpgtd %xmm6,%xmm14
1784 movdqu 32(%r12),%xmm9
1788 pshufd $0x13,%xmm14,%xmm13
1791 movdqa %xmm6,64(%rsp)
1794 pcmpgtd %xmm6,%xmm14
1796 movdqu 48(%r12),%xmm10
1800 pshufd $0x13,%xmm14,%xmm13
1803 movdqa %xmm6,80(%rsp)
1806 pcmpgtd %xmm6,%xmm14
1808 movdqu 64(%r12),%xmm11
1812 pshufd $0x13,%xmm14,%xmm13
1815 movdqa %xmm6,96(%rsp)
1818 pcmpgtd %xmm6,%xmm14
1820 movdqu 80(%r12),%xmm12
1824 movdqu 96(%r12),%xmm13
1826 movdqa %xmm6,112(%rsp)
1832 call _bsaes_encrypt8
1836 movdqu %xmm15,0(%r13)
1838 movdqu %xmm0,16(%r13)
1840 movdqu %xmm3,32(%r13)
1842 movdqu %xmm5,48(%r13)
1844 movdqu %xmm2,64(%r13)
1846 movdqu %xmm6,80(%r13)
1847 movdqu %xmm1,96(%r13)
1850 movdqa 112(%rsp),%xmm6
1860 call _bsaes_encrypt8
1864 movdqu %xmm15,0(%r13)
1866 movdqu %xmm0,16(%r13)
1868 movdqu %xmm3,32(%r13)
1870 movdqu %xmm5,48(%r13)
1872 movdqu %xmm2,64(%r13)
1873 movdqu %xmm6,80(%r13)
1876 movdqa 96(%rsp),%xmm6
1886 call _bsaes_encrypt8
1890 movdqu %xmm15,0(%r13)
1892 movdqu %xmm0,16(%r13)
1894 movdqu %xmm3,32(%r13)
1896 movdqu %xmm5,48(%r13)
1897 movdqu %xmm2,64(%r13)
1900 movdqa 80(%rsp),%xmm6
1910 call _bsaes_encrypt8
1914 movdqu %xmm15,0(%r13)
1916 movdqu %xmm0,16(%r13)
1918 movdqu %xmm3,32(%r13)
1919 movdqu %xmm5,48(%r13)
1922 movdqa 64(%rsp),%xmm6
1932 call _bsaes_encrypt8
1936 movdqu %xmm15,0(%r13)
1938 movdqu %xmm0,16(%r13)
1939 movdqu %xmm3,32(%r13)
1942 movdqa 48(%rsp),%xmm6
1952 call _bsaes_encrypt8
1956 movdqu %xmm15,0(%r13)
1957 movdqu %xmm0,16(%r13)
1960 movdqa 32(%rsp),%xmm6
1966 movdqa %xmm7,32(%rbp)
1970 call asm_AES_encrypt
1971 pxor 32(%rbp),%xmm15
1977 movdqu %xmm15,0(%r13)
1980 movdqa 16(%rsp),%xmm6
1989 movzbl -16(%rdx),%ecx
1997 movdqu -16(%r13),%xmm15
2001 movdqa %xmm15,32(%rbp)
2003 call asm_AES_encrypt
2005 movdqu %xmm6,-16(%r13)
2011 movdqa %xmm0,0(%rax)
2012 movdqa %xmm0,16(%rax)
2032 .cfi_def_cfa_register %rsp
2036 .size bsaes_xts_encrypt,.-bsaes_xts_encrypt
2038 .globl bsaes_xts_decrypt
2039 .type bsaes_xts_decrypt,@function
2046 .cfi_adjust_cfa_offset 8
2047 .cfi_offset %rbp,-16
2049 .cfi_adjust_cfa_offset 8
2050 .cfi_offset %rbx,-24
2052 .cfi_adjust_cfa_offset 8
2053 .cfi_offset %r12,-32
2055 .cfi_adjust_cfa_offset 8
2056 .cfi_offset %r13,-40
2058 .cfi_adjust_cfa_offset 8
2059 .cfi_offset %r14,-48
2061 .cfi_adjust_cfa_offset 8
2062 .cfi_offset %r15,-56
2064 .cfi_adjust_cfa_offset 0x48
2074 call asm_AES_encrypt
2087 call _bsaes_key_convert
2100 movdqa 32(%rbp),%xmm6
2103 movdqa .Lxts_magic(%rip),%xmm12
2104 pcmpgtd %xmm6,%xmm14
2112 pshufd $0x13,%xmm14,%xmm13
2115 movdqa %xmm6,0(%rsp)
2118 pcmpgtd %xmm6,%xmm14
2120 pshufd $0x13,%xmm14,%xmm13
2123 movdqa %xmm6,16(%rsp)
2126 pcmpgtd %xmm6,%xmm14
2128 movdqu 0(%r12),%xmm7
2129 pshufd $0x13,%xmm14,%xmm13
2132 movdqa %xmm6,32(%rsp)
2135 pcmpgtd %xmm6,%xmm14
2137 movdqu 16(%r12),%xmm8
2139 pshufd $0x13,%xmm14,%xmm13
2142 movdqa %xmm6,48(%rsp)
2145 pcmpgtd %xmm6,%xmm14
2147 movdqu 32(%r12),%xmm9
2149 pshufd $0x13,%xmm14,%xmm13
2152 movdqa %xmm6,64(%rsp)
2155 pcmpgtd %xmm6,%xmm14
2157 movdqu 48(%r12),%xmm10
2159 pshufd $0x13,%xmm14,%xmm13
2162 movdqa %xmm6,80(%rsp)
2165 pcmpgtd %xmm6,%xmm14
2167 movdqu 64(%r12),%xmm11
2169 pshufd $0x13,%xmm14,%xmm13
2172 movdqa %xmm6,96(%rsp)
2175 pcmpgtd %xmm6,%xmm14
2177 movdqu 80(%r12),%xmm12
2179 movdqu 96(%r12),%xmm13
2181 movdqu 112(%r12),%xmm14
2183 movdqa %xmm6,112(%rsp)
2189 call _bsaes_decrypt8
2193 movdqu %xmm15,0(%r13)
2195 movdqu %xmm0,16(%r13)
2197 movdqu %xmm5,32(%r13)
2199 movdqu %xmm3,48(%r13)
2201 movdqu %xmm1,64(%r13)
2203 movdqu %xmm6,80(%r13)
2204 pxor 112(%rsp),%xmm4
2205 movdqu %xmm2,96(%r13)
2206 movdqu %xmm4,112(%r13)
2209 movdqa 112(%rsp),%xmm6
2211 movdqa .Lxts_magic(%rip),%xmm12
2212 pcmpgtd %xmm6,%xmm14
2213 pshufd $0x13,%xmm14,%xmm13
2217 pcmpgtd %xmm6,%xmm14
2226 pshufd $0x13,%xmm14,%xmm13
2229 movdqa %xmm6,0(%rsp)
2232 pcmpgtd %xmm6,%xmm14
2234 pshufd $0x13,%xmm14,%xmm13
2237 movdqa %xmm6,16(%rsp)
2240 pcmpgtd %xmm6,%xmm14
2242 movdqu 0(%r12),%xmm7
2245 pshufd $0x13,%xmm14,%xmm13
2248 movdqa %xmm6,32(%rsp)
2251 pcmpgtd %xmm6,%xmm14
2253 movdqu 16(%r12),%xmm8
2257 pshufd $0x13,%xmm14,%xmm13
2260 movdqa %xmm6,48(%rsp)
2263 pcmpgtd %xmm6,%xmm14
2265 movdqu 32(%r12),%xmm9
2269 pshufd $0x13,%xmm14,%xmm13
2272 movdqa %xmm6,64(%rsp)
2275 pcmpgtd %xmm6,%xmm14
2277 movdqu 48(%r12),%xmm10
2281 pshufd $0x13,%xmm14,%xmm13
2284 movdqa %xmm6,80(%rsp)
2287 pcmpgtd %xmm6,%xmm14
2289 movdqu 64(%r12),%xmm11
2293 pshufd $0x13,%xmm14,%xmm13
2296 movdqa %xmm6,96(%rsp)
2299 pcmpgtd %xmm6,%xmm14
2301 movdqu 80(%r12),%xmm12
2305 movdqu 96(%r12),%xmm13
2307 movdqa %xmm6,112(%rsp)
2313 call _bsaes_decrypt8
2317 movdqu %xmm15,0(%r13)
2319 movdqu %xmm0,16(%r13)
2321 movdqu %xmm5,32(%r13)
2323 movdqu %xmm3,48(%r13)
2325 movdqu %xmm1,64(%r13)
2327 movdqu %xmm6,80(%r13)
2328 movdqu %xmm2,96(%r13)
2331 movdqa 112(%rsp),%xmm6
2341 call _bsaes_decrypt8
2345 movdqu %xmm15,0(%r13)
2347 movdqu %xmm0,16(%r13)
2349 movdqu %xmm5,32(%r13)
2351 movdqu %xmm3,48(%r13)
2353 movdqu %xmm1,64(%r13)
2354 movdqu %xmm6,80(%r13)
2357 movdqa 96(%rsp),%xmm6
2367 call _bsaes_decrypt8
2371 movdqu %xmm15,0(%r13)
2373 movdqu %xmm0,16(%r13)
2375 movdqu %xmm5,32(%r13)
2377 movdqu %xmm3,48(%r13)
2378 movdqu %xmm1,64(%r13)
2381 movdqa 80(%rsp),%xmm6
2391 call _bsaes_decrypt8
2395 movdqu %xmm15,0(%r13)
2397 movdqu %xmm0,16(%r13)
2399 movdqu %xmm5,32(%r13)
2400 movdqu %xmm3,48(%r13)
2403 movdqa 64(%rsp),%xmm6
2413 call _bsaes_decrypt8
2417 movdqu %xmm15,0(%r13)
2419 movdqu %xmm0,16(%r13)
2420 movdqu %xmm5,32(%r13)
2423 movdqa 48(%rsp),%xmm6
2433 call _bsaes_decrypt8
2437 movdqu %xmm15,0(%r13)
2438 movdqu %xmm0,16(%r13)
2441 movdqa 32(%rsp),%xmm6
2447 movdqa %xmm7,32(%rbp)
2451 call asm_AES_decrypt
2452 pxor 32(%rbp),%xmm15
2458 movdqu %xmm15,0(%r13)
2461 movdqa 16(%rsp),%xmm6
2468 movdqa .Lxts_magic(%rip),%xmm12
2469 pcmpgtd %xmm6,%xmm14
2470 pshufd $0x13,%xmm14,%xmm13
2474 movdqu (%r12),%xmm15
2480 movdqa %xmm15,32(%rbp)
2482 call asm_AES_decrypt
2488 movzbl 16(%r12),%eax
2497 movdqu (%r13),%xmm15
2501 movdqa %xmm15,32(%rbp)
2503 call asm_AES_decrypt
2511 movdqa %xmm0,0(%rax)
2512 movdqa %xmm0,16(%rax)
2532 .cfi_def_cfa_register %rsp
2536 .size bsaes_xts_decrypt,.-bsaes_xts_decrypt
2537 .type _bsaes_const,@object
2541 .quad 0x0a0e0206070b0f03, 0x0004080c0d010509
2543 .quad 0x01040b0e0205080f, 0x0306090c00070a0d
2545 .quad 0x0504070602010003, 0x0f0e0d0c080b0a09
2547 .quad 0x5555555555555555, 0x5555555555555555
2549 .quad 0x3333333333333333, 0x3333333333333333
2551 .quad 0x0f0f0f0f0f0f0f0f, 0x0f0f0f0f0f0f0f0f
2553 .quad 0x0504070600030201, 0x0f0e0d0c0a09080b
2555 .quad 0x0304090e00050a0f, 0x01060b0c0207080d
2557 .quad 0x0a0e02060f03070b, 0x0004080c05090d01
2559 .quad 0x0706050403020100, 0x0c0d0e0f0b0a0908
2561 .quad 0x0a0d02060c03070b, 0x0004080f05090e01
2563 .quad 0x0000000000000000, 0x0000000100000000
2565 .quad 0x0000000000000000, 0x0000000200000000
2567 .quad 0x0000000000000000, 0x0000000300000000
2569 .quad 0x0000000000000000, 0x0000000400000000
2571 .quad 0x0000000000000000, 0x0000000500000000
2573 .quad 0x0000000000000000, 0x0000000600000000
2575 .quad 0x0000000000000000, 0x0000000700000000
2577 .quad 0x0000000000000000, 0x0000000800000000
2581 .quad 0x0101010101010101, 0x0101010101010101
2582 .quad 0x0202020202020202, 0x0202020202020202
2583 .quad 0x0404040404040404, 0x0404040404040404
2584 .quad 0x0808080808080808, 0x0808080808080808
2586 .quad 0x02060a0e03070b0f, 0x0004080c0105090d
2588 .quad 0x6363636363636363, 0x6363636363636363
2589 .byte 66,105,116,45,115,108,105,99,101,100,32,65,69,83,32,102,111,114,32,120,56,54,95,54,52,47,83,83,83,69,51,44,32,69,109,105,108,105,97,32,75,195,164,115,112,101,114,44,32,80,101,116,101,114,32,83,99,104,119,97,98,101,44,32,65,110,100,121,32,80,111,108,121,97,107,111,118,0
2591 .size _bsaes_const,.-_bsaes_const