2 /* Do not modify. This file is auto-generated from rsaz-x86_64.pl. */
8 .type rsaz_512_sqr,@function
13 .cfi_adjust_cfa_offset 8
16 .cfi_adjust_cfa_offset 8
19 .cfi_adjust_cfa_offset 8
22 .cfi_adjust_cfa_offset 8
25 .cfi_adjust_cfa_offset 8
28 .cfi_adjust_cfa_offset 8
32 .cfi_adjust_cfa_offset 128+24
34 .byte 102,72,15,110,202
382 .byte 102,72,15,126,205
387 call __rsaz_512_reduce
399 call __rsaz_512_subtract
403 movl 128+8(%rsp),%r8d
409 leaq 128+24+48(%rsp),%rax
424 .cfi_def_cfa_register %rsp
428 .size rsaz_512_sqr,.-rsaz_512_sqr
430 .type rsaz_512_mul,@function
435 .cfi_adjust_cfa_offset 8
438 .cfi_adjust_cfa_offset 8
441 .cfi_adjust_cfa_offset 8
444 .cfi_adjust_cfa_offset 8
447 .cfi_adjust_cfa_offset 8
450 .cfi_adjust_cfa_offset 8
454 .cfi_adjust_cfa_offset 128+24
456 .byte 102,72,15,110,199
457 .byte 102,72,15,110,201
463 .byte 102,72,15,126,199
464 .byte 102,72,15,126,205
475 call __rsaz_512_reduce
486 call __rsaz_512_subtract
488 leaq 128+24+48(%rsp),%rax
503 .cfi_def_cfa_register %rsp
507 .size rsaz_512_mul,.-rsaz_512_mul
508 .globl rsaz_512_mul_gather4
509 .type rsaz_512_mul_gather4,@function
511 rsaz_512_mul_gather4:
514 .cfi_adjust_cfa_offset 8
517 .cfi_adjust_cfa_offset 8
520 .cfi_adjust_cfa_offset 8
523 .cfi_adjust_cfa_offset 8
526 .cfi_adjust_cfa_offset 8
529 .cfi_adjust_cfa_offset 8
533 .cfi_adjust_cfa_offset 152
536 movdqa .Linc+16(%rip),%xmm1
537 movdqa .Linc(%rip),%xmm0
539 pshufd $0,%xmm8,%xmm8
563 movdqa 16(%rdx),%xmm9
564 movdqa 32(%rdx),%xmm10
565 movdqa 48(%rdx),%xmm11
567 movdqa 64(%rdx),%xmm12
569 movdqa 80(%rdx),%xmm13
571 movdqa 96(%rdx),%xmm14
573 movdqa 112(%rdx),%xmm15
587 pshufd $0x4e,%xmm8,%xmm9
589 .byte 102,76,15,126,195
592 movq %rdi,128+8(%rsp)
593 movq %rcx,128+16(%rsp)
651 movdqa 16(%rbp),%xmm9
652 movdqa 32(%rbp),%xmm10
653 movdqa 48(%rbp),%xmm11
655 movdqa 64(%rbp),%xmm12
657 movdqa 80(%rbp),%xmm13
659 movdqa 96(%rbp),%xmm14
661 movdqa 112(%rbp),%xmm15
675 pshufd $0x4e,%xmm8,%xmm9
677 .byte 102,76,15,126,195
756 movq 128+8(%rsp),%rdi
757 movq 128+16(%rsp),%rbp
768 call __rsaz_512_reduce
779 call __rsaz_512_subtract
781 leaq 128+24+48(%rsp),%rax
796 .cfi_def_cfa_register %rsp
797 .Lmul_gather4_epilogue:
800 .size rsaz_512_mul_gather4,.-rsaz_512_mul_gather4
801 .globl rsaz_512_mul_scatter4
802 .type rsaz_512_mul_scatter4,@function
804 rsaz_512_mul_scatter4:
807 .cfi_adjust_cfa_offset 8
810 .cfi_adjust_cfa_offset 8
813 .cfi_adjust_cfa_offset 8
816 .cfi_adjust_cfa_offset 8
819 .cfi_adjust_cfa_offset 8
822 .cfi_adjust_cfa_offset 8
827 .cfi_adjust_cfa_offset 128+24
830 .byte 102,72,15,110,199
831 .byte 102,72,15,110,202
832 .byte 102,73,15,110,208
839 .byte 102,72,15,126,199
840 .byte 102,72,15,126,205
851 call __rsaz_512_reduce
860 .byte 102,72,15,126,214
863 call __rsaz_512_subtract
874 leaq 128+24+48(%rsp),%rax
889 .cfi_def_cfa_register %rsp
890 .Lmul_scatter4_epilogue:
893 .size rsaz_512_mul_scatter4,.-rsaz_512_mul_scatter4
894 .globl rsaz_512_mul_by_one
895 .type rsaz_512_mul_by_one,@function
900 .cfi_adjust_cfa_offset 8
903 .cfi_adjust_cfa_offset 8
906 .cfi_adjust_cfa_offset 8
909 .cfi_adjust_cfa_offset 8
912 .cfi_adjust_cfa_offset 8
915 .cfi_adjust_cfa_offset 8
919 .cfi_adjust_cfa_offset 128+24
935 movdqa %xmm0,16(%rsp)
936 movdqa %xmm0,32(%rsp)
937 movdqa %xmm0,48(%rsp)
938 movdqa %xmm0,64(%rsp)
939 movdqa %xmm0,80(%rsp)
940 movdqa %xmm0,96(%rsp)
941 call __rsaz_512_reduce
951 leaq 128+24+48(%rsp),%rax
966 .cfi_def_cfa_register %rsp
967 .Lmul_by_one_epilogue:
970 .size rsaz_512_mul_by_one,.-rsaz_512_mul_by_one
971 .type __rsaz_512_reduce,@function
976 imulq 128+8(%rsp),%rbx
1010 movq 128+8(%rsp),%rsi
1051 jne .Lreduction_loop
1055 .size __rsaz_512_reduce,.-__rsaz_512_reduce
1056 .type __rsaz_512_subtract,@function
1058 __rsaz_512_subtract:
1114 .size __rsaz_512_subtract,.-__rsaz_512_subtract
1115 .type __rsaz_512_mul,@function
1258 .size __rsaz_512_mul,.-__rsaz_512_mul
1259 .globl rsaz_512_scatter4
1260 .type rsaz_512_scatter4,@function
1264 leaq (%rdi,%rdx,8),%rdi
1277 .size rsaz_512_scatter4,.-rsaz_512_scatter4
1279 .globl rsaz_512_gather4
1280 .type rsaz_512_gather4,@function
1285 movdqa .Linc+16(%rip),%xmm1
1286 movdqa .Linc(%rip),%xmm0
1288 pshufd $0,%xmm8,%xmm8
1314 movdqa 0(%rsi),%xmm8
1315 movdqa 16(%rsi),%xmm9
1316 movdqa 32(%rsi),%xmm10
1317 movdqa 48(%rsi),%xmm11
1319 movdqa 64(%rsi),%xmm12
1321 movdqa 80(%rsi),%xmm13
1323 movdqa 96(%rsi),%xmm14
1325 movdqa 112(%rsi),%xmm15
1339 pshufd $0x4e,%xmm8,%xmm9
1346 .LSEH_end_rsaz_512_gather4:
1348 .size rsaz_512_gather4,.-rsaz_512_gather4