.text .file "" .globl makeNIST_P192L .align 16, 0x90 .type makeNIST_P192L,@function makeNIST_P192L: # @makeNIST_P192L # BB#0: movq $-1, %rax movq $-2, %rdx movq $-1, %rcx retq .Lfunc_end0: .size makeNIST_P192L, .Lfunc_end0-makeNIST_P192L .globl mcl_fpDbl_mod_NIST_P192L .align 16, 0x90 .type mcl_fpDbl_mod_NIST_P192L,@function mcl_fpDbl_mod_NIST_P192L: # @mcl_fpDbl_mod_NIST_P192L # BB#0: pushq %r14 pushq %rbx movq 16(%rsi), %r10 movq 24(%rsi), %r8 movq 40(%rsi), %r9 movq 8(%rsi), %rax addq %r9, %rax adcq $0, %r10 sbbq %rcx, %rcx andl $1, %ecx movq 32(%rsi), %r11 movq (%rsi), %r14 addq %r8, %r14 adcq %r11, %rax adcq %r9, %r10 adcq $0, %rcx addq %r9, %r14 adcq %r8, %rax adcq %r11, %r10 adcq $0, %rcx addq %rcx, %r14 adcq %rax, %rcx adcq $0, %r10 sbbq %rax, %rax andl $1, %eax movq %r14, %rsi addq $1, %rsi movq %rcx, %rdx adcq $1, %rdx movq %r10, %rbx adcq $0, %rbx adcq $-1, %rax andl $1, %eax cmovneq %r14, %rsi movq %rsi, (%rdi) testb %al, %al cmovneq %rcx, %rdx movq %rdx, 8(%rdi) cmovneq %r10, %rbx movq %rbx, 16(%rdi) popq %rbx popq %r14 retq .Lfunc_end1: .size mcl_fpDbl_mod_NIST_P192L, .Lfunc_end1-mcl_fpDbl_mod_NIST_P192L .globl mcl_fp_sqr_NIST_P192L .align 16, 0x90 .type mcl_fp_sqr_NIST_P192L,@function mcl_fp_sqr_NIST_P192L: # @mcl_fp_sqr_NIST_P192L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdi, -8(%rsp) # 8-byte Spill movq 16(%rsi), %r11 movq (%rsi), %rbx movq 8(%rsi), %rcx movq %r11, %rax mulq %rcx movq %rdx, %rdi movq %rax, %r14 movq %rcx, %rax mulq %rcx movq %rdx, %r15 movq %rax, %r12 movq %rcx, %rax mulq %rbx movq %rax, %r13 movq %rdx, %rcx addq %rcx, %r12 adcq %r14, %r15 movq %rdi, %r10 adcq $0, %r10 movq %r11, %rax mulq %rbx movq %rdx, %r9 movq %rax, %rbp movq %rbx, %rax mulq %rbx movq %rax, %r8 movq %rdx, %rsi addq %r13, %rsi adcq %rbp, %rcx movq %r9, %rbx adcq $0, %rbx addq %r13, %rsi adcq %r12, %rcx adcq %r15, %rbx adcq $0, %r10 movq %r11, %rax mulq %r11 addq %r14, %r9 adcq %rdi, %rax adcq $0, %rdx addq %rbp, %rcx adcq %rbx, %r9 adcq %r10, %rax adcq $0, %rdx addq %rdx, %rsi adcq $0, %rcx sbbq %rbp, %rbp andl $1, %ebp addq %r9, %r8 adcq %rax, %rsi adcq %rdx, %rcx adcq $0, %rbp addq %rdx, %r8 adcq %r9, %rsi adcq %rax, %rcx adcq $0, %rbp addq %rbp, %r8 adcq %rsi, %rbp adcq $0, %rcx sbbq %rax, %rax andl $1, %eax movq %r8, %rdx addq $1, %rdx movq %rbp, %rsi adcq $1, %rsi movq %rcx, %rdi adcq $0, %rdi adcq $-1, %rax andl $1, %eax cmovneq %r8, %rdx movq -8(%rsp), %rbx # 8-byte Reload movq %rdx, (%rbx) testb %al, %al cmovneq %rbp, %rsi movq %rsi, 8(%rbx) cmovneq %rcx, %rdi movq %rdi, 16(%rbx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end2: .size mcl_fp_sqr_NIST_P192L, .Lfunc_end2-mcl_fp_sqr_NIST_P192L .globl mcl_fp_mulNIST_P192L .align 16, 0x90 .type mcl_fp_mulNIST_P192L,@function mcl_fp_mulNIST_P192L: # @mcl_fp_mulNIST_P192L # BB#0: pushq %r14 pushq %rbx subq $56, %rsp movq %rdi, %r14 leaq 8(%rsp), %rdi callq mcl_fpDbl_mulPre3L@PLT movq 24(%rsp), %r9 movq 32(%rsp), %r8 movq 48(%rsp), %rdi movq 16(%rsp), %rbx addq %rdi, %rbx adcq $0, %r9 sbbq %rcx, %rcx andl $1, %ecx movq 40(%rsp), %rsi movq 8(%rsp), %rdx addq %r8, %rdx adcq %rsi, %rbx adcq %rdi, %r9 adcq $0, %rcx addq %rdi, %rdx adcq %r8, %rbx adcq %rsi, %r9 adcq $0, %rcx addq %rcx, %rdx adcq %rbx, %rcx adcq $0, %r9 sbbq %rsi, %rsi andl $1, %esi movq %rdx, %rdi addq $1, %rdi movq %rcx, %rbx adcq $1, %rbx movq %r9, %rax adcq $0, %rax adcq $-1, %rsi andl $1, %esi cmovneq %rdx, %rdi movq %rdi, (%r14) testb %sil, %sil cmovneq %rcx, %rbx movq %rbx, 8(%r14) cmovneq %r9, %rax movq %rax, 16(%r14) addq $56, %rsp popq %rbx popq %r14 retq .Lfunc_end3: .size mcl_fp_mulNIST_P192L, .Lfunc_end3-mcl_fp_mulNIST_P192L .globl mcl_fpDbl_mod_NIST_P521L .align 16, 0x90 .type mcl_fpDbl_mod_NIST_P521L,@function mcl_fpDbl_mod_NIST_P521L: # @mcl_fpDbl_mod_NIST_P521L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq 120(%rsi), %r9 movq 128(%rsi), %r14 movq %r14, %r8 shldq $55, %r9, %r8 movq 112(%rsi), %r10 shldq $55, %r10, %r9 movq 104(%rsi), %r11 shldq $55, %r11, %r10 movq 96(%rsi), %r15 shldq $55, %r15, %r11 movq 88(%rsi), %r12 shldq $55, %r12, %r15 movq 80(%rsi), %rcx shldq $55, %rcx, %r12 movq 64(%rsi), %rbx movq 72(%rsi), %rax shldq $55, %rax, %rcx shrq $9, %r14 shldq $55, %rbx, %rax andl $511, %ebx # imm = 0x1FF addq (%rsi), %rax adcq 8(%rsi), %rcx adcq 16(%rsi), %r12 adcq 24(%rsi), %r15 adcq 32(%rsi), %r11 adcq 40(%rsi), %r10 adcq 48(%rsi), %r9 adcq 56(%rsi), %r8 adcq %r14, %rbx movq %rbx, %rsi shrq $9, %rsi andl $1, %esi addq %rax, %rsi adcq $0, %rcx adcq $0, %r12 adcq $0, %r15 adcq $0, %r11 adcq $0, %r10 adcq $0, %r9 adcq $0, %r8 adcq $0, %rbx movq %rsi, %rax andq %r12, %rax andq %r15, %rax andq %r11, %rax andq %r10, %rax andq %r9, %rax andq %r8, %rax movq %rbx, %rdx orq $-512, %rdx # imm = 0xFFFFFFFFFFFFFE00 andq %rax, %rdx andq %rcx, %rdx cmpq $-1, %rdx je .LBB4_1 # BB#3: # %nonzero movq %rsi, (%rdi) movq %rcx, 8(%rdi) movq %r12, 16(%rdi) movq %r15, 24(%rdi) movq %r11, 32(%rdi) movq %r10, 40(%rdi) movq %r9, 48(%rdi) movq %r8, 56(%rdi) andl $511, %ebx # imm = 0x1FF movq %rbx, 64(%rdi) jmp .LBB4_2 .LBB4_1: # %zero movq $0, 64(%rdi) movq $0, 56(%rdi) movq $0, 48(%rdi) movq $0, 40(%rdi) movq $0, 32(%rdi) movq $0, 24(%rdi) movq $0, 16(%rdi) movq $0, 8(%rdi) movq $0, (%rdi) .LBB4_2: # %zero popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end4: .size mcl_fpDbl_mod_NIST_P521L, .Lfunc_end4-mcl_fpDbl_mod_NIST_P521L .globl mcl_fp_mulUnitPre1L .align 16, 0x90 .type mcl_fp_mulUnitPre1L,@function mcl_fp_mulUnitPre1L: # @mcl_fp_mulUnitPre1L # BB#0: movq %rdx, %rax mulq (%rsi) movq %rax, (%rdi) movq %rdx, 8(%rdi) retq .Lfunc_end5: .size mcl_fp_mulUnitPre1L, .Lfunc_end5-mcl_fp_mulUnitPre1L .globl mcl_fpDbl_mulPre1L .align 16, 0x90 .type mcl_fpDbl_mulPre1L,@function mcl_fpDbl_mulPre1L: # @mcl_fpDbl_mulPre1L # BB#0: movq (%rdx), %rax mulq (%rsi) movq %rax, (%rdi) movq %rdx, 8(%rdi) retq .Lfunc_end6: .size mcl_fpDbl_mulPre1L, .Lfunc_end6-mcl_fpDbl_mulPre1L .globl mcl_fpDbl_sqrPre1L .align 16, 0x90 .type mcl_fpDbl_sqrPre1L,@function mcl_fpDbl_sqrPre1L: # @mcl_fpDbl_sqrPre1L # BB#0: movq (%rsi), %rax mulq %rax movq %rax, (%rdi) movq %rdx, 8(%rdi) retq .Lfunc_end7: .size mcl_fpDbl_sqrPre1L, .Lfunc_end7-mcl_fpDbl_sqrPre1L .globl mcl_fp_mont1L .align 16, 0x90 .type mcl_fp_mont1L,@function mcl_fp_mont1L: # @mcl_fp_mont1L # BB#0: movq (%rsi), %rax mulq (%rdx) movq %rax, %rsi movq %rdx, %r8 movq -8(%rcx), %rax imulq %rsi, %rax movq (%rcx), %rcx mulq %rcx addq %rsi, %rax adcq %r8, %rdx sbbq %rax, %rax andl $1, %eax movq %rdx, %rsi subq %rcx, %rsi sbbq $0, %rax testb $1, %al cmovneq %rdx, %rsi movq %rsi, (%rdi) retq .Lfunc_end8: .size mcl_fp_mont1L, .Lfunc_end8-mcl_fp_mont1L .globl mcl_fp_montNF1L .align 16, 0x90 .type mcl_fp_montNF1L,@function mcl_fp_montNF1L: # @mcl_fp_montNF1L # BB#0: movq (%rsi), %rax mulq (%rdx) movq %rax, %rsi movq %rdx, %r8 movq -8(%rcx), %rax imulq %rsi, %rax movq (%rcx), %rcx mulq %rcx addq %rsi, %rax adcq %r8, %rdx movq %rdx, %rax subq %rcx, %rax cmovsq %rdx, %rax movq %rax, (%rdi) retq .Lfunc_end9: .size mcl_fp_montNF1L, .Lfunc_end9-mcl_fp_montNF1L .globl mcl_fp_montRed1L .align 16, 0x90 .type mcl_fp_montRed1L,@function mcl_fp_montRed1L: # @mcl_fp_montRed1L # BB#0: movq (%rsi), %rcx movq -8(%rdx), %rax imulq %rcx, %rax movq (%rdx), %r8 mulq %r8 addq %rcx, %rax adcq 8(%rsi), %rdx sbbq %rax, %rax andl $1, %eax movq %rdx, %rcx subq %r8, %rcx sbbq $0, %rax testb $1, %al cmovneq %rdx, %rcx movq %rcx, (%rdi) retq .Lfunc_end10: .size mcl_fp_montRed1L, .Lfunc_end10-mcl_fp_montRed1L .globl mcl_fp_addPre1L .align 16, 0x90 .type mcl_fp_addPre1L,@function mcl_fp_addPre1L: # @mcl_fp_addPre1L # BB#0: movq (%rdx), %rax addq (%rsi), %rax movq %rax, (%rdi) sbbq %rax, %rax andl $1, %eax retq .Lfunc_end11: .size mcl_fp_addPre1L, .Lfunc_end11-mcl_fp_addPre1L .globl mcl_fp_subPre1L .align 16, 0x90 .type mcl_fp_subPre1L,@function mcl_fp_subPre1L: # @mcl_fp_subPre1L # BB#0: movq (%rsi), %rcx xorl %eax, %eax subq (%rdx), %rcx movq %rcx, (%rdi) sbbq $0, %rax andl $1, %eax retq .Lfunc_end12: .size mcl_fp_subPre1L, .Lfunc_end12-mcl_fp_subPre1L .globl mcl_fp_shr1_1L .align 16, 0x90 .type mcl_fp_shr1_1L,@function mcl_fp_shr1_1L: # @mcl_fp_shr1_1L # BB#0: movq (%rsi), %rax shrq %rax movq %rax, (%rdi) retq .Lfunc_end13: .size mcl_fp_shr1_1L, .Lfunc_end13-mcl_fp_shr1_1L .globl mcl_fp_add1L .align 16, 0x90 .type mcl_fp_add1L,@function mcl_fp_add1L: # @mcl_fp_add1L # BB#0: movq (%rdx), %rax addq (%rsi), %rax movq %rax, (%rdi) sbbq %rdx, %rdx andl $1, %edx subq (%rcx), %rax sbbq $0, %rdx testb $1, %dl jne .LBB14_2 # BB#1: # %nocarry movq %rax, (%rdi) .LBB14_2: # %carry retq .Lfunc_end14: .size mcl_fp_add1L, .Lfunc_end14-mcl_fp_add1L .globl mcl_fp_addNF1L .align 16, 0x90 .type mcl_fp_addNF1L,@function mcl_fp_addNF1L: # @mcl_fp_addNF1L # BB#0: movq (%rdx), %rax addq (%rsi), %rax movq %rax, %rdx subq (%rcx), %rdx cmovsq %rax, %rdx movq %rdx, (%rdi) retq .Lfunc_end15: .size mcl_fp_addNF1L, .Lfunc_end15-mcl_fp_addNF1L .globl mcl_fp_sub1L .align 16, 0x90 .type mcl_fp_sub1L,@function mcl_fp_sub1L: # @mcl_fp_sub1L # BB#0: movq (%rsi), %rax xorl %esi, %esi subq (%rdx), %rax movq %rax, (%rdi) sbbq $0, %rsi testb $1, %sil jne .LBB16_2 # BB#1: # %nocarry retq .LBB16_2: # %carry addq (%rcx), %rax movq %rax, (%rdi) retq .Lfunc_end16: .size mcl_fp_sub1L, .Lfunc_end16-mcl_fp_sub1L .globl mcl_fp_subNF1L .align 16, 0x90 .type mcl_fp_subNF1L,@function mcl_fp_subNF1L: # @mcl_fp_subNF1L # BB#0: movq (%rsi), %rax subq (%rdx), %rax movq %rax, %rdx sarq $63, %rdx andq (%rcx), %rdx addq %rax, %rdx movq %rdx, (%rdi) retq .Lfunc_end17: .size mcl_fp_subNF1L, .Lfunc_end17-mcl_fp_subNF1L .globl mcl_fpDbl_add1L .align 16, 0x90 .type mcl_fpDbl_add1L,@function mcl_fpDbl_add1L: # @mcl_fpDbl_add1L # BB#0: movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx movq %rax, (%rdi) sbbq %rax, %rax andl $1, %eax movq %rdx, %rsi subq (%rcx), %rsi sbbq $0, %rax testb $1, %al cmovneq %rdx, %rsi movq %rsi, 8(%rdi) retq .Lfunc_end18: .size mcl_fpDbl_add1L, .Lfunc_end18-mcl_fpDbl_add1L .globl mcl_fpDbl_sub1L .align 16, 0x90 .type mcl_fpDbl_sub1L,@function mcl_fpDbl_sub1L: # @mcl_fpDbl_sub1L # BB#0: movq (%rsi), %rax movq 8(%rsi), %r8 xorl %esi, %esi subq (%rdx), %rax sbbq 8(%rdx), %r8 movq %rax, (%rdi) movl $0, %eax sbbq $0, %rax testb $1, %al cmovneq (%rcx), %rsi addq %r8, %rsi movq %rsi, 8(%rdi) retq .Lfunc_end19: .size mcl_fpDbl_sub1L, .Lfunc_end19-mcl_fpDbl_sub1L .globl mcl_fp_mulUnitPre2L .align 16, 0x90 .type mcl_fp_mulUnitPre2L,@function mcl_fp_mulUnitPre2L: # @mcl_fp_mulUnitPre2L # BB#0: movq %rdx, %r8 movq %r8, %rax mulq 8(%rsi) movq %rdx, %rcx movq %rax, %r9 movq %r8, %rax mulq (%rsi) movq %rax, (%rdi) addq %r9, %rdx movq %rdx, 8(%rdi) adcq $0, %rcx movq %rcx, 16(%rdi) retq .Lfunc_end20: .size mcl_fp_mulUnitPre2L, .Lfunc_end20-mcl_fp_mulUnitPre2L .globl mcl_fpDbl_mulPre2L .align 16, 0x90 .type mcl_fpDbl_mulPre2L,@function mcl_fpDbl_mulPre2L: # @mcl_fpDbl_mulPre2L # BB#0: pushq %r14 pushq %rbx movq %rdx, %r10 movq (%rsi), %r8 movq 8(%rsi), %r11 movq (%r10), %rcx movq %r8, %rax mulq %rcx movq %rdx, %r9 movq %rax, (%rdi) movq %r11, %rax mulq %rcx movq %rdx, %r14 movq %rax, %rsi addq %r9, %rsi adcq $0, %r14 movq 8(%r10), %rbx movq %r11, %rax mulq %rbx movq %rdx, %r9 movq %rax, %rcx movq %r8, %rax mulq %rbx addq %rsi, %rax movq %rax, 8(%rdi) adcq %r14, %rcx sbbq %rax, %rax andl $1, %eax addq %rdx, %rcx movq %rcx, 16(%rdi) adcq %r9, %rax movq %rax, 24(%rdi) popq %rbx popq %r14 retq .Lfunc_end21: .size mcl_fpDbl_mulPre2L, .Lfunc_end21-mcl_fpDbl_mulPre2L .globl mcl_fpDbl_sqrPre2L .align 16, 0x90 .type mcl_fpDbl_sqrPre2L,@function mcl_fpDbl_sqrPre2L: # @mcl_fpDbl_sqrPre2L # BB#0: movq (%rsi), %rcx movq 8(%rsi), %r8 movq %rcx, %rax mulq %rcx movq %rdx, %rsi movq %rax, (%rdi) movq %r8, %rax mulq %rcx movq %rdx, %r9 movq %rax, %r10 addq %r10, %rsi movq %r9, %rcx adcq $0, %rcx movq %r8, %rax mulq %r8 addq %r10, %rsi movq %rsi, 8(%rdi) adcq %rcx, %rax sbbq %rcx, %rcx andl $1, %ecx addq %r9, %rax movq %rax, 16(%rdi) adcq %rdx, %rcx movq %rcx, 24(%rdi) retq .Lfunc_end22: .size mcl_fpDbl_sqrPre2L, .Lfunc_end22-mcl_fpDbl_sqrPre2L .globl mcl_fp_mont2L .align 16, 0x90 .type mcl_fp_mont2L,@function mcl_fp_mont2L: # @mcl_fp_mont2L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdi, -8(%rsp) # 8-byte Spill movq (%rsi), %r8 movq 8(%rsi), %r11 movq (%rdx), %rsi movq 8(%rdx), %r9 movq %r11, %rax mulq %rsi movq %rdx, %r15 movq %rax, %r10 movq %r8, %rax mulq %rsi movq %rax, %r14 movq %rdx, %r13 addq %r10, %r13 adcq $0, %r15 movq -8(%rcx), %r10 movq (%rcx), %rbp movq %r14, %rsi imulq %r10, %rsi movq 8(%rcx), %rdi movq %rsi, %rax mulq %rdi movq %rdx, %rcx movq %rax, %r12 movq %rsi, %rax mulq %rbp movq %rdx, %rbx addq %r12, %rbx adcq $0, %rcx addq %r14, %rax adcq %r13, %rbx adcq %r15, %rcx sbbq %r15, %r15 andl $1, %r15d movq %r9, %rax mulq %r11 movq %rdx, %r14 movq %rax, %r11 movq %r9, %rax mulq %r8 movq %rax, %r8 movq %rdx, %rsi addq %r11, %rsi adcq $0, %r14 addq %rbx, %r8 adcq %rcx, %rsi adcq %r15, %r14 sbbq %rbx, %rbx andl $1, %ebx imulq %r8, %r10 movq %r10, %rax mulq %rdi movq %rdx, %rcx movq %rax, %r9 movq %r10, %rax mulq %rbp addq %r9, %rdx adcq $0, %rcx addq %r8, %rax adcq %rsi, %rdx adcq %r14, %rcx adcq $0, %rbx movq %rdx, %rax subq %rbp, %rax movq %rcx, %rsi sbbq %rdi, %rsi sbbq $0, %rbx andl $1, %ebx cmovneq %rcx, %rsi testb %bl, %bl cmovneq %rdx, %rax movq -8(%rsp), %rcx # 8-byte Reload movq %rax, (%rcx) movq %rsi, 8(%rcx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end23: .size mcl_fp_mont2L, .Lfunc_end23-mcl_fp_mont2L .globl mcl_fp_montNF2L .align 16, 0x90 .type mcl_fp_montNF2L,@function mcl_fp_montNF2L: # @mcl_fp_montNF2L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdi, -8(%rsp) # 8-byte Spill movq (%rsi), %r8 movq 8(%rsi), %r11 movq (%rdx), %rbp movq 8(%rdx), %r9 movq %r8, %rax mulq %rbp movq %rax, %rsi movq %rdx, %r14 movq -8(%rcx), %r10 movq (%rcx), %r15 movq %rsi, %rbx imulq %r10, %rbx movq 8(%rcx), %rdi movq %rbx, %rax mulq %rdi movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %r13 movq %rbx, %rax mulq %r15 movq %rdx, %r12 movq %rax, %rbx movq %r11, %rax mulq %rbp movq %rdx, %rcx movq %rax, %rbp addq %r14, %rbp adcq $0, %rcx addq %rsi, %rbx adcq %r13, %rbp adcq $0, %rcx addq %r12, %rbp adcq -16(%rsp), %rcx # 8-byte Folded Reload movq %r9, %rax mulq %r11 movq %rdx, %rsi movq %rax, %r11 movq %r9, %rax mulq %r8 movq %rax, %r8 movq %rdx, %rbx addq %r11, %rbx adcq $0, %rsi addq %rbp, %r8 adcq %rcx, %rbx adcq $0, %rsi imulq %r8, %r10 movq %r10, %rax mulq %rdi movq %rdx, %rcx movq %rax, %rbp movq %r10, %rax mulq %r15 addq %r8, %rax adcq %rbx, %rbp adcq $0, %rsi addq %rdx, %rbp adcq %rcx, %rsi movq %rbp, %rax subq %r15, %rax movq %rsi, %rcx sbbq %rdi, %rcx cmovsq %rbp, %rax movq -8(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovsq %rsi, %rcx movq %rcx, 8(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end24: .size mcl_fp_montNF2L, .Lfunc_end24-mcl_fp_montNF2L .globl mcl_fp_montRed2L .align 16, 0x90 .type mcl_fp_montRed2L,@function mcl_fp_montRed2L: # @mcl_fp_montRed2L # BB#0: pushq %r15 pushq %r14 pushq %rbx movq -8(%rdx), %r9 movq (%rdx), %r11 movq (%rsi), %rbx movq %rbx, %rcx imulq %r9, %rcx movq 8(%rdx), %r14 movq %rcx, %rax mulq %r14 movq %rdx, %r8 movq %rax, %r10 movq %rcx, %rax mulq %r11 movq %rdx, %rcx addq %r10, %rcx adcq $0, %r8 movq 24(%rsi), %r15 addq %rbx, %rax adcq 8(%rsi), %rcx adcq 16(%rsi), %r8 adcq $0, %r15 sbbq %rbx, %rbx andl $1, %ebx imulq %rcx, %r9 movq %r9, %rax mulq %r14 movq %rdx, %rsi movq %rax, %r10 movq %r9, %rax mulq %r11 addq %r10, %rdx adcq $0, %rsi addq %rcx, %rax adcq %r8, %rdx adcq %r15, %rsi adcq $0, %rbx movq %rdx, %rax subq %r11, %rax movq %rsi, %rcx sbbq %r14, %rcx sbbq $0, %rbx andl $1, %ebx cmovneq %rsi, %rcx testb %bl, %bl cmovneq %rdx, %rax movq %rax, (%rdi) movq %rcx, 8(%rdi) popq %rbx popq %r14 popq %r15 retq .Lfunc_end25: .size mcl_fp_montRed2L, .Lfunc_end25-mcl_fp_montRed2L .globl mcl_fp_addPre2L .align 16, 0x90 .type mcl_fp_addPre2L,@function mcl_fp_addPre2L: # @mcl_fp_addPre2L # BB#0: movq (%rdx), %rax movq 8(%rdx), %rcx addq (%rsi), %rax adcq 8(%rsi), %rcx movq %rax, (%rdi) movq %rcx, 8(%rdi) sbbq %rax, %rax andl $1, %eax retq .Lfunc_end26: .size mcl_fp_addPre2L, .Lfunc_end26-mcl_fp_addPre2L .globl mcl_fp_subPre2L .align 16, 0x90 .type mcl_fp_subPre2L,@function mcl_fp_subPre2L: # @mcl_fp_subPre2L # BB#0: movq (%rsi), %rcx movq 8(%rsi), %rsi xorl %eax, %eax subq (%rdx), %rcx sbbq 8(%rdx), %rsi movq %rcx, (%rdi) movq %rsi, 8(%rdi) sbbq $0, %rax andl $1, %eax retq .Lfunc_end27: .size mcl_fp_subPre2L, .Lfunc_end27-mcl_fp_subPre2L .globl mcl_fp_shr1_2L .align 16, 0x90 .type mcl_fp_shr1_2L,@function mcl_fp_shr1_2L: # @mcl_fp_shr1_2L # BB#0: movq (%rsi), %rax movq 8(%rsi), %rcx shrdq $1, %rcx, %rax movq %rax, (%rdi) shrq %rcx movq %rcx, 8(%rdi) retq .Lfunc_end28: .size mcl_fp_shr1_2L, .Lfunc_end28-mcl_fp_shr1_2L .globl mcl_fp_add2L .align 16, 0x90 .type mcl_fp_add2L,@function mcl_fp_add2L: # @mcl_fp_add2L # BB#0: movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx movq %rax, (%rdi) movq %rdx, 8(%rdi) sbbq %rsi, %rsi andl $1, %esi subq (%rcx), %rax sbbq 8(%rcx), %rdx sbbq $0, %rsi testb $1, %sil jne .LBB29_2 # BB#1: # %nocarry movq %rax, (%rdi) movq %rdx, 8(%rdi) .LBB29_2: # %carry retq .Lfunc_end29: .size mcl_fp_add2L, .Lfunc_end29-mcl_fp_add2L .globl mcl_fp_addNF2L .align 16, 0x90 .type mcl_fp_addNF2L,@function mcl_fp_addNF2L: # @mcl_fp_addNF2L # BB#0: movq (%rdx), %rax movq 8(%rdx), %r8 addq (%rsi), %rax adcq 8(%rsi), %r8 movq %rax, %rsi subq (%rcx), %rsi movq %r8, %rdx sbbq 8(%rcx), %rdx testq %rdx, %rdx cmovsq %rax, %rsi movq %rsi, (%rdi) cmovsq %r8, %rdx movq %rdx, 8(%rdi) retq .Lfunc_end30: .size mcl_fp_addNF2L, .Lfunc_end30-mcl_fp_addNF2L .globl mcl_fp_sub2L .align 16, 0x90 .type mcl_fp_sub2L,@function mcl_fp_sub2L: # @mcl_fp_sub2L # BB#0: movq (%rsi), %rax movq 8(%rsi), %r8 xorl %esi, %esi subq (%rdx), %rax sbbq 8(%rdx), %r8 movq %rax, (%rdi) movq %r8, 8(%rdi) sbbq $0, %rsi testb $1, %sil jne .LBB31_2 # BB#1: # %nocarry retq .LBB31_2: # %carry movq 8(%rcx), %rdx addq (%rcx), %rax movq %rax, (%rdi) adcq %r8, %rdx movq %rdx, 8(%rdi) retq .Lfunc_end31: .size mcl_fp_sub2L, .Lfunc_end31-mcl_fp_sub2L .globl mcl_fp_subNF2L .align 16, 0x90 .type mcl_fp_subNF2L,@function mcl_fp_subNF2L: # @mcl_fp_subNF2L # BB#0: movq (%rsi), %r8 movq 8(%rsi), %rsi subq (%rdx), %r8 sbbq 8(%rdx), %rsi movq %rsi, %rdx sarq $63, %rdx movq 8(%rcx), %rax andq %rdx, %rax andq (%rcx), %rdx addq %r8, %rdx movq %rdx, (%rdi) adcq %rsi, %rax movq %rax, 8(%rdi) retq .Lfunc_end32: .size mcl_fp_subNF2L, .Lfunc_end32-mcl_fp_subNF2L .globl mcl_fpDbl_add2L .align 16, 0x90 .type mcl_fpDbl_add2L,@function mcl_fpDbl_add2L: # @mcl_fpDbl_add2L # BB#0: movq 24(%rdx), %r8 movq 24(%rsi), %r9 movq 16(%rdx), %r10 movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %r10 movq %rax, (%rdi) movq %rdx, 8(%rdi) adcq %r8, %r9 sbbq %rax, %rax andl $1, %eax movq %r10, %rdx subq (%rcx), %rdx movq %r9, %rsi sbbq 8(%rcx), %rsi sbbq $0, %rax andl $1, %eax cmovneq %r10, %rdx movq %rdx, 16(%rdi) testb %al, %al cmovneq %r9, %rsi movq %rsi, 24(%rdi) retq .Lfunc_end33: .size mcl_fpDbl_add2L, .Lfunc_end33-mcl_fpDbl_add2L .globl mcl_fpDbl_sub2L .align 16, 0x90 .type mcl_fpDbl_sub2L,@function mcl_fpDbl_sub2L: # @mcl_fpDbl_sub2L # BB#0: movq 24(%rdx), %r8 movq 24(%rsi), %r9 movq 16(%rsi), %r10 movq (%rsi), %r11 movq 8(%rsi), %rsi xorl %eax, %eax subq (%rdx), %r11 sbbq 8(%rdx), %rsi sbbq 16(%rdx), %r10 movq %r11, (%rdi) movq %rsi, 8(%rdi) sbbq %r8, %r9 movl $0, %edx sbbq $0, %rdx andl $1, %edx movq (%rcx), %rsi cmoveq %rax, %rsi testb %dl, %dl cmovneq 8(%rcx), %rax addq %r10, %rsi movq %rsi, 16(%rdi) adcq %r9, %rax movq %rax, 24(%rdi) retq .Lfunc_end34: .size mcl_fpDbl_sub2L, .Lfunc_end34-mcl_fpDbl_sub2L .globl mcl_fp_mulUnitPre3L .align 16, 0x90 .type mcl_fp_mulUnitPre3L,@function mcl_fp_mulUnitPre3L: # @mcl_fp_mulUnitPre3L # BB#0: movq %rdx, %rcx movq %rcx, %rax mulq 16(%rsi) movq %rdx, %r8 movq %rax, %r9 movq %rcx, %rax mulq 8(%rsi) movq %rdx, %r10 movq %rax, %r11 movq %rcx, %rax mulq (%rsi) movq %rax, (%rdi) addq %r11, %rdx movq %rdx, 8(%rdi) adcq %r9, %r10 movq %r10, 16(%rdi) adcq $0, %r8 movq %r8, 24(%rdi) retq .Lfunc_end35: .size mcl_fp_mulUnitPre3L, .Lfunc_end35-mcl_fp_mulUnitPre3L .globl mcl_fpDbl_mulPre3L .align 16, 0x90 .type mcl_fpDbl_mulPre3L,@function mcl_fpDbl_mulPre3L: # @mcl_fpDbl_mulPre3L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %r10 movq (%rsi), %r8 movq 8(%rsi), %r9 movq (%r10), %rbx movq %r8, %rax mulq %rbx movq %rdx, %rcx movq 16(%rsi), %r11 movq %rax, (%rdi) movq %r11, %rax mulq %rbx movq %rdx, %r14 movq %rax, %rsi movq %r9, %rax mulq %rbx movq %rdx, %r15 movq %rax, %rbx addq %rcx, %rbx adcq %rsi, %r15 adcq $0, %r14 movq 8(%r10), %rcx movq %r11, %rax mulq %rcx movq %rdx, %r12 movq %rax, %rbp movq %r9, %rax mulq %rcx movq %rdx, %r13 movq %rax, %rsi movq %r8, %rax mulq %rcx addq %rbx, %rax movq %rax, 8(%rdi) adcq %r15, %rsi adcq %r14, %rbp sbbq %r14, %r14 andl $1, %r14d addq %rdx, %rsi adcq %r13, %rbp adcq %r12, %r14 movq 16(%r10), %r15 movq %r11, %rax mulq %r15 movq %rdx, %r10 movq %rax, %rbx movq %r9, %rax mulq %r15 movq %rdx, %r9 movq %rax, %rcx movq %r8, %rax mulq %r15 addq %rsi, %rax movq %rax, 16(%rdi) adcq %rbp, %rcx adcq %r14, %rbx sbbq %rax, %rax andl $1, %eax addq %rdx, %rcx movq %rcx, 24(%rdi) adcq %r9, %rbx movq %rbx, 32(%rdi) adcq %r10, %rax movq %rax, 40(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end36: .size mcl_fpDbl_mulPre3L, .Lfunc_end36-mcl_fpDbl_mulPre3L .globl mcl_fpDbl_sqrPre3L .align 16, 0x90 .type mcl_fpDbl_sqrPre3L,@function mcl_fpDbl_sqrPre3L: # @mcl_fpDbl_sqrPre3L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 16(%rsi), %r10 movq (%rsi), %rcx movq 8(%rsi), %rsi movq %rcx, %rax mulq %rcx movq %rdx, %rbx movq %rax, (%rdi) movq %r10, %rax mulq %rcx movq %rdx, %r8 movq %rax, %r11 movq %rsi, %rax mulq %rcx movq %rdx, %r14 movq %rax, %r12 addq %r12, %rbx movq %r14, %r13 adcq %r11, %r13 movq %r8, %rcx adcq $0, %rcx movq %r10, %rax mulq %rsi movq %rdx, %r9 movq %rax, %r15 movq %rsi, %rax mulq %rsi movq %rax, %rsi addq %r12, %rbx movq %rbx, 8(%rdi) adcq %r13, %rsi adcq %r15, %rcx sbbq %rbx, %rbx andl $1, %ebx addq %r14, %rsi adcq %rdx, %rcx adcq %r9, %rbx movq %r10, %rax mulq %r10 addq %r11, %rsi movq %rsi, 16(%rdi) adcq %r15, %rcx adcq %rbx, %rax sbbq %rsi, %rsi andl $1, %esi addq %r8, %rcx movq %rcx, 24(%rdi) adcq %r9, %rax movq %rax, 32(%rdi) adcq %rdx, %rsi movq %rsi, 40(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end37: .size mcl_fpDbl_sqrPre3L, .Lfunc_end37-mcl_fpDbl_sqrPre3L .globl mcl_fp_mont3L .align 16, 0x90 .type mcl_fp_mont3L,@function mcl_fp_mont3L: # @mcl_fp_mont3L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %r10 movq %r10, -56(%rsp) # 8-byte Spill movq %rdi, -48(%rsp) # 8-byte Spill movq 16(%rsi), %rax movq %rax, -64(%rsp) # 8-byte Spill movq (%r10), %rdi mulq %rdi movq %rax, %rbp movq %rdx, %r8 movq (%rsi), %rbx movq %rbx, -32(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -40(%rsp) # 8-byte Spill mulq %rdi movq %rdx, %r15 movq %rax, %rsi movq %rbx, %rax mulq %rdi movq %rax, %r12 movq %rdx, %r11 addq %rsi, %r11 adcq %rbp, %r15 adcq $0, %r8 movq -8(%rcx), %r14 movq (%rcx), %rdi movq %rdi, -24(%rsp) # 8-byte Spill movq %r12, %rbp imulq %r14, %rbp movq 16(%rcx), %rdx movq %rdx, -16(%rsp) # 8-byte Spill movq 8(%rcx), %rbx movq %rbx, -8(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rdx, %rcx movq %rax, %r13 movq %rbp, %rax mulq %rbx movq %rdx, %rsi movq %rax, %r9 movq %rbp, %rax mulq %rdi movq %rdx, %rbp addq %r9, %rbp adcq %r13, %rsi adcq $0, %rcx addq %r12, %rax adcq %r11, %rbp movq 8(%r10), %rbx adcq %r15, %rsi adcq %r8, %rcx sbbq %rdi, %rdi andl $1, %edi movq %rbx, %rax movq -64(%rsp), %r10 # 8-byte Reload mulq %r10 movq %rdx, %r15 movq %rax, %r9 movq %rbx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %r11 movq %rbx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r8 movq %rdx, %rbx addq %r11, %rbx adcq %r9, %r12 adcq $0, %r15 addq %rbp, %r8 adcq %rsi, %rbx adcq %rcx, %r12 adcq %rdi, %r15 sbbq %r11, %r11 andl $1, %r11d movq %r8, %rcx imulq %r14, %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r9 movq %rcx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %rdi movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbp addq %rdi, %rbp adcq %r9, %rsi adcq $0, %r13 addq %r8, %rax adcq %rbx, %rbp adcq %r12, %rsi adcq %r15, %r13 adcq $0, %r11 movq -56(%rsp), %rax # 8-byte Reload movq 16(%rax), %rcx movq %rcx, %rax mulq %r10 movq %rdx, %r8 movq %rax, %r10 movq %rcx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %rdi movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r9 movq %rdx, %rcx addq %rdi, %rcx adcq %r10, %r15 adcq $0, %r8 addq %rbp, %r9 adcq %rsi, %rcx adcq %r13, %r15 adcq %r11, %r8 sbbq %rdi, %rdi andl $1, %edi imulq %r9, %r14 movq %r14, %rax movq -16(%rsp), %r12 # 8-byte Reload mulq %r12 movq %rdx, %rbx movq %rax, %r10 movq %r14, %rax movq -8(%rsp), %r13 # 8-byte Reload mulq %r13 movq %rdx, %rsi movq %rax, %r11 movq %r14, %rax movq -24(%rsp), %rbp # 8-byte Reload mulq %rbp addq %r11, %rdx adcq %r10, %rsi adcq $0, %rbx addq %r9, %rax adcq %rcx, %rdx adcq %r15, %rsi adcq %r8, %rbx adcq $0, %rdi movq %rdx, %rax subq %rbp, %rax movq %rsi, %rcx sbbq %r13, %rcx movq %rbx, %rbp sbbq %r12, %rbp sbbq $0, %rdi andl $1, %edi cmovneq %rbx, %rbp testb %dil, %dil cmovneq %rdx, %rax movq -48(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovneq %rsi, %rcx movq %rcx, 8(%rdx) movq %rbp, 16(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end38: .size mcl_fp_mont3L, .Lfunc_end38-mcl_fp_mont3L .globl mcl_fp_montNF3L .align 16, 0x90 .type mcl_fp_montNF3L,@function mcl_fp_montNF3L: # @mcl_fp_montNF3L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -8(%rsp) # 8-byte Spill movq %rdi, -32(%rsp) # 8-byte Spill movq 16(%rsi), %r10 movq %r10, -40(%rsp) # 8-byte Spill movq (%rdx), %rbp movq %r10, %rax mulq %rbp movq %rax, %r14 movq %rdx, %r15 movq (%rsi), %rbx movq %rbx, -64(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -24(%rsp) # 8-byte Spill mulq %rbp movq %rdx, %rdi movq %rax, %rsi movq %rbx, %rax mulq %rbp movq %rax, %r13 movq %rdx, %rbp addq %rsi, %rbp adcq %r14, %rdi adcq $0, %r15 movq -8(%rcx), %r14 movq (%rcx), %r11 movq %r11, -48(%rsp) # 8-byte Spill movq %r13, %rbx imulq %r14, %rbx movq 16(%rcx), %rdx movq %rdx, -16(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -56(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rdx movq %rdx, %r8 movq %rax, %r12 movq %rbx, %rax mulq %rcx movq %rdx, %r9 movq %rax, %rcx movq %rbx, %rax mulq %r11 addq %r13, %rax adcq %rbp, %rcx adcq %rdi, %r12 adcq $0, %r15 addq %rdx, %rcx movq -8(%rsp), %rax # 8-byte Reload movq 8(%rax), %rbp adcq %r9, %r12 adcq %r8, %r15 movq %rbp, %rax mulq %r10 movq %rdx, %rsi movq %rax, %r8 movq %rbp, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r9 movq %rbp, %rax movq -64(%rsp), %r10 # 8-byte Reload mulq %r10 movq %rax, %r13 movq %rdx, %rbp addq %r9, %rbp adcq %r8, %rbx adcq $0, %rsi addq %rcx, %r13 adcq %r12, %rbp adcq %r15, %rbx adcq $0, %rsi movq %r13, %rcx imulq %r14, %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r15 movq %rcx, %rax movq -56(%rsp), %rdi # 8-byte Reload mulq %rdi movq %rdx, %r9 movq %rax, %r12 movq %rcx, %rax mulq %r11 addq %r13, %rax adcq %rbp, %r12 adcq %rbx, %r15 adcq $0, %rsi addq %rdx, %r12 adcq %r9, %r15 adcq %r8, %rsi movq -8(%rsp), %rax # 8-byte Reload movq 16(%rax), %rbx movq %rbx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r8 movq %rbx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r9 movq %rbx, %rax mulq %r10 movq %rax, %r10 movq %rdx, %rbx addq %r9, %rbx adcq %r8, %rcx adcq $0, %rbp addq %r12, %r10 adcq %r15, %rbx adcq %rsi, %rcx adcq $0, %rbp imulq %r10, %r14 movq %r14, %rax movq -16(%rsp), %r15 # 8-byte Reload mulq %r15 movq %rdx, %r8 movq %rax, %rsi movq %r14, %rax movq %rdi, %r11 mulq %r11 movq %rdx, %r9 movq %rax, %rdi movq %r14, %rax movq -48(%rsp), %r14 # 8-byte Reload mulq %r14 addq %r10, %rax adcq %rbx, %rdi adcq %rcx, %rsi adcq $0, %rbp addq %rdx, %rdi adcq %r9, %rsi adcq %r8, %rbp movq %rdi, %rax subq %r14, %rax movq %rsi, %rcx sbbq %r11, %rcx movq %rbp, %rbx sbbq %r15, %rbx movq %rbx, %rdx sarq $63, %rdx cmovsq %rdi, %rax movq -32(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovsq %rsi, %rcx movq %rcx, 8(%rdx) cmovsq %rbp, %rbx movq %rbx, 16(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end39: .size mcl_fp_montNF3L, .Lfunc_end39-mcl_fp_montNF3L .globl mcl_fp_montRed3L .align 16, 0x90 .type mcl_fp_montRed3L,@function mcl_fp_montRed3L: # @mcl_fp_montRed3L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rdi, -8(%rsp) # 8-byte Spill movq -8(%rcx), %r9 movq (%rcx), %rdi movq %rdi, -16(%rsp) # 8-byte Spill movq (%rsi), %r15 movq %r15, %rbx imulq %r9, %rbx movq 16(%rcx), %rbp movq %rbp, -24(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rbp movq %rax, %r11 movq %rdx, %r8 movq 8(%rcx), %rcx movq %rcx, -32(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rcx movq %rcx, %r12 movq %rdx, %r10 movq %rax, %r14 movq %rbx, %rax mulq %rdi movq %rdi, %rbx movq %rdx, %rcx addq %r14, %rcx adcq %r11, %r10 adcq $0, %r8 movq 40(%rsi), %rdi movq 32(%rsi), %r13 addq %r15, %rax adcq 8(%rsi), %rcx adcq 16(%rsi), %r10 adcq 24(%rsi), %r8 adcq $0, %r13 adcq $0, %rdi sbbq %r15, %r15 andl $1, %r15d movq %rcx, %rsi imulq %r9, %rsi movq %rsi, %rax mulq %rbp movq %rdx, %r11 movq %rax, %rbp movq %rsi, %rax mulq %r12 movq %rdx, %r14 movq %rax, %r12 movq %rsi, %rax mulq %rbx movq %rdx, %rbx addq %r12, %rbx adcq %rbp, %r14 adcq $0, %r11 addq %rcx, %rax adcq %r10, %rbx adcq %r8, %r14 adcq %r13, %r11 adcq $0, %rdi adcq $0, %r15 imulq %rbx, %r9 movq %r9, %rax movq -24(%rsp), %r12 # 8-byte Reload mulq %r12 movq %rdx, %rbp movq %rax, %r8 movq %r9, %rax movq -32(%rsp), %r13 # 8-byte Reload mulq %r13 movq %rdx, %rsi movq %rax, %r10 movq %r9, %rax movq -16(%rsp), %rcx # 8-byte Reload mulq %rcx addq %r10, %rdx adcq %r8, %rsi adcq $0, %rbp addq %rbx, %rax adcq %r14, %rdx adcq %r11, %rsi adcq %rdi, %rbp adcq $0, %r15 movq %rdx, %rax subq %rcx, %rax movq %rsi, %rdi sbbq %r13, %rdi movq %rbp, %rcx sbbq %r12, %rcx sbbq $0, %r15 andl $1, %r15d cmovneq %rbp, %rcx testb %r15b, %r15b cmovneq %rdx, %rax movq -8(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovneq %rsi, %rdi movq %rdi, 8(%rdx) movq %rcx, 16(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end40: .size mcl_fp_montRed3L, .Lfunc_end40-mcl_fp_montRed3L .globl mcl_fp_addPre3L .align 16, 0x90 .type mcl_fp_addPre3L,@function mcl_fp_addPre3L: # @mcl_fp_addPre3L # BB#0: movq 16(%rdx), %rax movq (%rdx), %rcx movq 8(%rdx), %rdx addq (%rsi), %rcx adcq 8(%rsi), %rdx adcq 16(%rsi), %rax movq %rcx, (%rdi) movq %rdx, 8(%rdi) movq %rax, 16(%rdi) sbbq %rax, %rax andl $1, %eax retq .Lfunc_end41: .size mcl_fp_addPre3L, .Lfunc_end41-mcl_fp_addPre3L .globl mcl_fp_subPre3L .align 16, 0x90 .type mcl_fp_subPre3L,@function mcl_fp_subPre3L: # @mcl_fp_subPre3L # BB#0: movq 16(%rsi), %r8 movq (%rsi), %rcx movq 8(%rsi), %rsi xorl %eax, %eax subq (%rdx), %rcx sbbq 8(%rdx), %rsi sbbq 16(%rdx), %r8 movq %rcx, (%rdi) movq %rsi, 8(%rdi) movq %r8, 16(%rdi) sbbq $0, %rax andl $1, %eax retq .Lfunc_end42: .size mcl_fp_subPre3L, .Lfunc_end42-mcl_fp_subPre3L .globl mcl_fp_shr1_3L .align 16, 0x90 .type mcl_fp_shr1_3L,@function mcl_fp_shr1_3L: # @mcl_fp_shr1_3L # BB#0: movq 16(%rsi), %rax movq (%rsi), %rcx movq 8(%rsi), %rdx shrdq $1, %rdx, %rcx movq %rcx, (%rdi) shrdq $1, %rax, %rdx movq %rdx, 8(%rdi) shrq %rax movq %rax, 16(%rdi) retq .Lfunc_end43: .size mcl_fp_shr1_3L, .Lfunc_end43-mcl_fp_shr1_3L .globl mcl_fp_add3L .align 16, 0x90 .type mcl_fp_add3L,@function mcl_fp_add3L: # @mcl_fp_add3L # BB#0: movq 16(%rdx), %r8 movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %r8 movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r8, 16(%rdi) sbbq %rsi, %rsi andl $1, %esi subq (%rcx), %rax sbbq 8(%rcx), %rdx sbbq 16(%rcx), %r8 sbbq $0, %rsi testb $1, %sil jne .LBB44_2 # BB#1: # %nocarry movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r8, 16(%rdi) .LBB44_2: # %carry retq .Lfunc_end44: .size mcl_fp_add3L, .Lfunc_end44-mcl_fp_add3L .globl mcl_fp_addNF3L .align 16, 0x90 .type mcl_fp_addNF3L,@function mcl_fp_addNF3L: # @mcl_fp_addNF3L # BB#0: movq 16(%rdx), %r8 movq (%rdx), %r10 movq 8(%rdx), %r9 addq (%rsi), %r10 adcq 8(%rsi), %r9 adcq 16(%rsi), %r8 movq %r10, %rsi subq (%rcx), %rsi movq %r9, %rdx sbbq 8(%rcx), %rdx movq %r8, %rax sbbq 16(%rcx), %rax movq %rax, %rcx sarq $63, %rcx cmovsq %r10, %rsi movq %rsi, (%rdi) cmovsq %r9, %rdx movq %rdx, 8(%rdi) cmovsq %r8, %rax movq %rax, 16(%rdi) retq .Lfunc_end45: .size mcl_fp_addNF3L, .Lfunc_end45-mcl_fp_addNF3L .globl mcl_fp_sub3L .align 16, 0x90 .type mcl_fp_sub3L,@function mcl_fp_sub3L: # @mcl_fp_sub3L # BB#0: movq 16(%rsi), %r8 movq (%rsi), %rax movq 8(%rsi), %r9 xorl %esi, %esi subq (%rdx), %rax sbbq 8(%rdx), %r9 sbbq 16(%rdx), %r8 movq %rax, (%rdi) movq %r9, 8(%rdi) movq %r8, 16(%rdi) sbbq $0, %rsi testb $1, %sil jne .LBB46_2 # BB#1: # %nocarry retq .LBB46_2: # %carry movq 8(%rcx), %rdx movq 16(%rcx), %rsi addq (%rcx), %rax movq %rax, (%rdi) adcq %r9, %rdx movq %rdx, 8(%rdi) adcq %r8, %rsi movq %rsi, 16(%rdi) retq .Lfunc_end46: .size mcl_fp_sub3L, .Lfunc_end46-mcl_fp_sub3L .globl mcl_fp_subNF3L .align 16, 0x90 .type mcl_fp_subNF3L,@function mcl_fp_subNF3L: # @mcl_fp_subNF3L # BB#0: movq 16(%rsi), %r10 movq (%rsi), %r8 movq 8(%rsi), %r9 subq (%rdx), %r8 sbbq 8(%rdx), %r9 sbbq 16(%rdx), %r10 movq %r10, %rdx sarq $63, %rdx movq %rdx, %rsi shldq $1, %r10, %rsi andq (%rcx), %rsi movq 16(%rcx), %rax andq %rdx, %rax andq 8(%rcx), %rdx addq %r8, %rsi movq %rsi, (%rdi) adcq %r9, %rdx movq %rdx, 8(%rdi) adcq %r10, %rax movq %rax, 16(%rdi) retq .Lfunc_end47: .size mcl_fp_subNF3L, .Lfunc_end47-mcl_fp_subNF3L .globl mcl_fpDbl_add3L .align 16, 0x90 .type mcl_fpDbl_add3L,@function mcl_fpDbl_add3L: # @mcl_fpDbl_add3L # BB#0: pushq %r15 pushq %r14 pushq %rbx movq 40(%rdx), %r10 movq 40(%rsi), %r8 movq 32(%rdx), %r11 movq 24(%rdx), %r14 movq 24(%rsi), %r15 movq 32(%rsi), %r9 movq 16(%rdx), %rbx movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %rbx movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %rbx, 16(%rdi) adcq %r14, %r15 adcq %r11, %r9 adcq %r10, %r8 sbbq %rax, %rax andl $1, %eax movq %r15, %rdx subq (%rcx), %rdx movq %r9, %rsi sbbq 8(%rcx), %rsi movq %r8, %rbx sbbq 16(%rcx), %rbx sbbq $0, %rax andl $1, %eax cmovneq %r15, %rdx movq %rdx, 24(%rdi) testb %al, %al cmovneq %r9, %rsi movq %rsi, 32(%rdi) cmovneq %r8, %rbx movq %rbx, 40(%rdi) popq %rbx popq %r14 popq %r15 retq .Lfunc_end48: .size mcl_fpDbl_add3L, .Lfunc_end48-mcl_fpDbl_add3L .globl mcl_fpDbl_sub3L .align 16, 0x90 .type mcl_fpDbl_sub3L,@function mcl_fpDbl_sub3L: # @mcl_fpDbl_sub3L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq 40(%rdx), %r10 movq 40(%rsi), %r8 movq 32(%rsi), %r9 movq 24(%rsi), %r11 movq 16(%rsi), %r14 movq (%rsi), %rbx movq 8(%rsi), %rax xorl %esi, %esi subq (%rdx), %rbx sbbq 8(%rdx), %rax movq 24(%rdx), %r15 movq 32(%rdx), %r12 sbbq 16(%rdx), %r14 movq %rbx, (%rdi) movq %rax, 8(%rdi) movq %r14, 16(%rdi) sbbq %r15, %r11 sbbq %r12, %r9 sbbq %r10, %r8 movl $0, %eax sbbq $0, %rax andl $1, %eax movq (%rcx), %rdx cmoveq %rsi, %rdx testb %al, %al movq 16(%rcx), %rax cmoveq %rsi, %rax cmovneq 8(%rcx), %rsi addq %r11, %rdx movq %rdx, 24(%rdi) adcq %r9, %rsi movq %rsi, 32(%rdi) adcq %r8, %rax movq %rax, 40(%rdi) popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end49: .size mcl_fpDbl_sub3L, .Lfunc_end49-mcl_fpDbl_sub3L .globl mcl_fp_mulUnitPre4L .align 16, 0x90 .type mcl_fp_mulUnitPre4L,@function mcl_fp_mulUnitPre4L: # @mcl_fp_mulUnitPre4L # BB#0: pushq %r14 pushq %rbx movq %rdx, %rcx movq %rcx, %rax mulq 24(%rsi) movq %rdx, %r8 movq %rax, %r9 movq %rcx, %rax mulq 16(%rsi) movq %rdx, %r10 movq %rax, %r11 movq %rcx, %rax mulq 8(%rsi) movq %rdx, %rbx movq %rax, %r14 movq %rcx, %rax mulq (%rsi) movq %rax, (%rdi) addq %r14, %rdx movq %rdx, 8(%rdi) adcq %r11, %rbx movq %rbx, 16(%rdi) adcq %r9, %r10 movq %r10, 24(%rdi) adcq $0, %r8 movq %r8, 32(%rdi) popq %rbx popq %r14 retq .Lfunc_end50: .size mcl_fp_mulUnitPre4L, .Lfunc_end50-mcl_fp_mulUnitPre4L .globl mcl_fpDbl_mulPre4L .align 16, 0x90 .type mcl_fpDbl_mulPre4L,@function mcl_fpDbl_mulPre4L: # @mcl_fpDbl_mulPre4L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -16(%rsp) # 8-byte Spill movq (%rsi), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 8(%rsi), %r8 movq %r8, -64(%rsp) # 8-byte Spill movq (%rdx), %rbx movq %rdx, %rbp mulq %rbx movq %rdx, %r15 movq 16(%rsi), %rcx movq %rcx, -24(%rsp) # 8-byte Spill movq 24(%rsi), %r11 movq %rax, (%rdi) movq %r11, %rax mulq %rbx movq %rdx, %r12 movq %rax, %r14 movq %rcx, %rax mulq %rbx movq %rdx, %r10 movq %rax, %r9 movq %r8, %rax mulq %rbx movq %rdx, %r13 movq %rax, %r8 addq %r15, %r8 adcq %r9, %r13 adcq %r14, %r10 adcq $0, %r12 movq %rbp, %r9 movq 8(%r9), %rbp movq %r11, %rax mulq %rbp movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r15 movq %rcx, %rax mulq %rbp movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %rcx movq -64(%rsp), %r14 # 8-byte Reload movq %r14, %rax mulq %rbp movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %rbx movq -8(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -56(%rsp) # 8-byte Spill addq %r8, %rax movq %rax, 8(%rdi) adcq %r13, %rbx adcq %r10, %rcx adcq %r12, %r15 sbbq %r13, %r13 andl $1, %r13d movq 16(%r9), %rbp movq %r14, %rax mulq %rbp movq %rax, %r12 movq %rdx, %r8 addq -56(%rsp), %rbx # 8-byte Folded Reload adcq -48(%rsp), %rcx # 8-byte Folded Reload adcq -40(%rsp), %r15 # 8-byte Folded Reload adcq -32(%rsp), %r13 # 8-byte Folded Reload movq %r11, %rax mulq %rbp movq %rdx, %r9 movq %rax, %r11 movq -24(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, %r14 movq %rax, %r10 movq -8(%rsp), %rax # 8-byte Reload mulq %rbp addq %rbx, %rax movq %rax, 16(%rdi) adcq %r12, %rcx adcq %r15, %r10 adcq %r13, %r11 sbbq %r13, %r13 andl $1, %r13d addq %rdx, %rcx adcq %r8, %r10 adcq %r14, %r11 adcq %r9, %r13 movq -16(%rsp), %rax # 8-byte Reload movq 24(%rax), %rbx movq %rbx, %rax mulq 24(%rsi) movq %rdx, %r8 movq %rax, %r14 movq %rbx, %rax mulq 16(%rsi) movq %rdx, %r9 movq %rax, %r12 movq %rbx, %rax mulq 8(%rsi) movq %rdx, %r15 movq %rax, %rbp movq %rbx, %rax mulq (%rsi) addq %rcx, %rax movq %rax, 24(%rdi) adcq %r10, %rbp adcq %r11, %r12 adcq %r13, %r14 sbbq %rax, %rax andl $1, %eax addq %rdx, %rbp movq %rbp, 32(%rdi) adcq %r15, %r12 movq %r12, 40(%rdi) adcq %r9, %r14 movq %r14, 48(%rdi) adcq %r8, %rax movq %rax, 56(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end51: .size mcl_fpDbl_mulPre4L, .Lfunc_end51-mcl_fpDbl_mulPre4L .globl mcl_fpDbl_sqrPre4L .align 16, 0x90 .type mcl_fpDbl_sqrPre4L,@function mcl_fpDbl_sqrPre4L: # @mcl_fpDbl_sqrPre4L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rsi, %r10 movq 16(%r10), %r9 movq 24(%r10), %r11 movq (%r10), %r15 movq 8(%r10), %r8 movq %r15, %rax mulq %r15 movq %rdx, %rbp movq %rax, (%rdi) movq %r11, %rax mulq %r8 movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, -32(%rsp) # 8-byte Spill movq %r9, %rax mulq %r8 movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, -40(%rsp) # 8-byte Spill movq %r11, %rax mulq %r15 movq %rdx, %rbx movq %rax, %rcx movq %r9, %rax mulq %r15 movq %rdx, %rsi movq %rsi, -16(%rsp) # 8-byte Spill movq %rax, %r12 movq %r8, %rax mulq %r8 movq %rdx, %r13 movq %rax, %r14 movq %r8, %rax mulq %r15 addq %rax, %rbp movq %rdx, %r8 adcq %r12, %r8 adcq %rsi, %rcx adcq $0, %rbx addq %rax, %rbp movq %rbp, 8(%rdi) adcq %r14, %r8 movq -40(%rsp), %rsi # 8-byte Reload adcq %rsi, %rcx adcq -32(%rsp), %rbx # 8-byte Folded Reload sbbq %rbp, %rbp andl $1, %ebp addq %rdx, %r8 adcq %r13, %rcx movq -24(%rsp), %r15 # 8-byte Reload adcq %r15, %rbx adcq -8(%rsp), %rbp # 8-byte Folded Reload movq %r11, %rax mulq %r9 movq %rdx, %r14 movq %rax, %r11 movq %r9, %rax mulq %r9 movq %rax, %r9 addq %r12, %r8 movq %r8, 16(%rdi) adcq %rsi, %rcx adcq %rbx, %r9 adcq %rbp, %r11 sbbq %r12, %r12 andl $1, %r12d addq -16(%rsp), %rcx # 8-byte Folded Reload adcq %r15, %r9 adcq %rdx, %r11 adcq %r14, %r12 movq 24(%r10), %rbp movq %rbp, %rax mulq 16(%r10) movq %rdx, %r8 movq %rax, %r14 movq %rbp, %rax mulq 8(%r10) movq %rdx, %r13 movq %rax, %rbx movq %rbp, %rax mulq (%r10) movq %rdx, %r15 movq %rax, %rsi movq %rbp, %rax mulq %rbp addq %rcx, %rsi movq %rsi, 24(%rdi) adcq %r9, %rbx adcq %r11, %r14 adcq %r12, %rax sbbq %rcx, %rcx andl $1, %ecx addq %r15, %rbx movq %rbx, 32(%rdi) adcq %r13, %r14 movq %r14, 40(%rdi) adcq %r8, %rax movq %rax, 48(%rdi) adcq %rdx, %rcx movq %rcx, 56(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end52: .size mcl_fpDbl_sqrPre4L, .Lfunc_end52-mcl_fpDbl_sqrPre4L .globl mcl_fp_mont4L .align 16, 0x90 .type mcl_fp_mont4L,@function mcl_fp_mont4L: # @mcl_fp_mont4L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -48(%rsp) # 8-byte Spill movq %rdi, -88(%rsp) # 8-byte Spill movq 24(%rsi), %rax movq %rax, -56(%rsp) # 8-byte Spill movq (%rdx), %rdi mulq %rdi movq %rax, %r9 movq %rdx, %rbp movq 16(%rsi), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rdi movq %rax, %r8 movq %rdx, %r10 movq (%rsi), %rbx movq %rbx, -72(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -80(%rsp) # 8-byte Spill mulq %rdi movq %rdx, %r14 movq %rax, %rsi movq %rbx, %rax mulq %rdi movq %rax, %r11 movq %rdx, %r13 addq %rsi, %r13 adcq %r8, %r14 adcq %r9, %r10 adcq $0, %rbp movq %rbp, -96(%rsp) # 8-byte Spill movq -8(%rcx), %rax movq %rax, -24(%rsp) # 8-byte Spill movq (%rcx), %r8 movq %r8, -32(%rsp) # 8-byte Spill movq %r11, %rdi imulq %rax, %rdi movq 24(%rcx), %rdx movq %rdx, -8(%rsp) # 8-byte Spill movq 16(%rcx), %rsi movq %rsi, -16(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -40(%rsp) # 8-byte Spill movq %rdi, %rax mulq %rdx movq %rdx, %r9 movq %rax, %r12 movq %rdi, %rax mulq %rsi movq %rdx, %rbp movq %rax, %rbx movq %rdi, %rax mulq %rcx movq %rdx, %rsi movq %rax, %r15 movq %rdi, %rax mulq %r8 movq %rdx, %rcx addq %r15, %rcx adcq %rbx, %rsi adcq %r12, %rbp adcq $0, %r9 addq %r11, %rax adcq %r13, %rcx adcq %r14, %rsi adcq %r10, %rbp adcq -96(%rsp), %r9 # 8-byte Folded Reload sbbq %r13, %r13 andl $1, %r13d movq -48(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdi movq %rdi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %r11 movq %rdi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r14 movq %rdi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r15 movq %rdi, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rax, %r8 movq %rdx, %rdi addq %r15, %rdi adcq %r14, %rbx adcq %r11, %r10 adcq $0, %r12 addq %rcx, %r8 adcq %rsi, %rdi adcq %rbp, %rbx adcq %r9, %r10 adcq %r13, %r12 sbbq %r15, %r15 andl $1, %r15d movq %r8, %rsi imulq -24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -96(%rsp) # 8-byte Spill movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r14 movq %rsi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %rbp movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r11 addq %rbp, %r11 adcq %r14, %r9 adcq -96(%rsp), %rcx # 8-byte Folded Reload adcq $0, %r13 addq %r8, %rax adcq %rdi, %r11 adcq %rbx, %r9 adcq %r10, %rcx adcq %r12, %r13 adcq $0, %r15 movq -48(%rsp), %rax # 8-byte Reload movq 16(%rax), %rsi movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -96(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rbx movq %rsi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %rdi movq %rsi, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rax, %r8 movq %rdx, %rbp addq %rdi, %rbp adcq %rbx, %r14 adcq -96(%rsp), %r10 # 8-byte Folded Reload adcq $0, %r12 addq %r11, %r8 adcq %r9, %rbp adcq %rcx, %r14 adcq %r13, %r10 adcq %r15, %r12 sbbq %r13, %r13 movq %r8, %rsi imulq -24(%rsp), %rsi # 8-byte Folded Reload andl $1, %r13d movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r9 movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r11 movq %rsi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r15 movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rsi addq %r15, %rsi adcq %r11, %rbx adcq %r9, %rcx adcq $0, %rdi addq %r8, %rax adcq %rbp, %rsi adcq %r14, %rbx adcq %r10, %rcx adcq %r12, %rdi adcq $0, %r13 movq -48(%rsp), %rax # 8-byte Reload movq 24(%rax), %rbp movq %rbp, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r14 movq %rbp, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r15 movq %rbp, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %r12 movq %rbp, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rax, %r9 movq %rdx, %rbp addq %r12, %rbp adcq %r15, %r11 adcq %r14, %r10 adcq $0, %r8 addq %rsi, %r9 adcq %rbx, %rbp adcq %rcx, %r11 adcq %rdi, %r10 adcq %r13, %r8 sbbq %rsi, %rsi andl $1, %esi movq -24(%rsp), %rcx # 8-byte Reload imulq %r9, %rcx movq %rcx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, -24(%rsp) # 8-byte Spill movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r15 movq %rcx, %rax movq -40(%rsp), %r14 # 8-byte Reload mulq %r14 movq %rdx, %rdi movq %rax, %r12 movq %rcx, %rax movq -32(%rsp), %rcx # 8-byte Reload mulq %rcx addq %r12, %rdx adcq %r15, %rdi adcq -24(%rsp), %r13 # 8-byte Folded Reload adcq $0, %rbx addq %r9, %rax adcq %rbp, %rdx adcq %r11, %rdi adcq %r10, %r13 adcq %r8, %rbx adcq $0, %rsi movq %rdx, %rax subq %rcx, %rax movq %rdi, %rcx sbbq %r14, %rcx movq %r13, %r8 sbbq -16(%rsp), %r8 # 8-byte Folded Reload movq %rbx, %rbp sbbq -8(%rsp), %rbp # 8-byte Folded Reload sbbq $0, %rsi andl $1, %esi cmovneq %rbx, %rbp testb %sil, %sil cmovneq %rdx, %rax movq -88(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovneq %rdi, %rcx movq %rcx, 8(%rdx) cmovneq %r13, %r8 movq %r8, 16(%rdx) movq %rbp, 24(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end53: .size mcl_fp_mont4L, .Lfunc_end53-mcl_fp_mont4L .globl mcl_fp_montNF4L .align 16, 0x90 .type mcl_fp_montNF4L,@function mcl_fp_montNF4L: # @mcl_fp_montNF4L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -16(%rsp) # 8-byte Spill movq %rdi, -88(%rsp) # 8-byte Spill movq 24(%rsi), %rax movq %rax, -24(%rsp) # 8-byte Spill movq (%rdx), %rdi mulq %rdi movq %rax, %r15 movq %rdx, %r12 movq 16(%rsi), %rax movq %rax, -32(%rsp) # 8-byte Spill mulq %rdi movq %rax, %r8 movq %rdx, %r9 movq (%rsi), %rbp movq %rbp, -40(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -48(%rsp) # 8-byte Spill mulq %rdi movq %rdx, %rbx movq %rax, %rsi movq %rbp, %rax mulq %rdi movq %rax, %r11 movq %rdx, %rdi addq %rsi, %rdi adcq %r8, %rbx adcq %r15, %r9 adcq $0, %r12 movq -8(%rcx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq (%rcx), %r8 movq %r8, -64(%rsp) # 8-byte Spill movq %r11, %rsi imulq %rax, %rsi movq 24(%rcx), %rdx movq %rdx, -56(%rsp) # 8-byte Spill movq 16(%rcx), %rbp movq %rbp, -72(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -80(%rsp) # 8-byte Spill movq %rsi, %rax mulq %rdx movq %rdx, %r15 movq %rax, %r13 movq %rsi, %rax mulq %rbp movq %rdx, %r10 movq %rax, %rbp movq %rsi, %rax mulq %rcx movq %rdx, %r14 movq %rax, %rcx movq %rsi, %rax mulq %r8 addq %r11, %rax adcq %rdi, %rcx adcq %rbx, %rbp adcq %r9, %r13 adcq $0, %r12 addq %rdx, %rcx adcq %r14, %rbp adcq %r10, %r13 adcq %r15, %r12 movq -16(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdi movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r10 movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r11 movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r14 movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rax, %rdi movq %rdx, %r9 addq %r14, %r9 adcq %r11, %r8 adcq %r10, %rsi adcq $0, %rbx addq %rcx, %rdi adcq %rbp, %r9 adcq %r13, %r8 adcq %r12, %rsi adcq $0, %rbx movq %rdi, %rcx imulq -8(%rsp), %rcx # 8-byte Folded Reload movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r12 movq %rcx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %r13 movq %rcx, %rax movq -80(%rsp), %r15 # 8-byte Reload mulq %r15 movq %rdx, %r14 movq %rax, %rbp movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload addq %rdi, %rax adcq %r9, %rbp adcq %r8, %r13 adcq %rsi, %r12 adcq $0, %rbx addq %rdx, %rbp adcq %r14, %r13 adcq %r11, %r12 adcq %r10, %rbx movq -16(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdi movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r10 movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r11 movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r14 movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rax, %r9 movq %rdx, %rdi addq %r14, %rdi adcq %r11, %rcx adcq %r10, %r8 adcq $0, %rsi addq %rbp, %r9 adcq %r13, %rdi adcq %r12, %rcx adcq %rbx, %r8 adcq $0, %rsi movq %r9, %rbx imulq -8(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r12 movq %rbx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %r13 movq %rbx, %rax mulq %r15 movq %rdx, %r14 movq %rax, %rbp movq %rbx, %rax mulq -64(%rsp) # 8-byte Folded Reload addq %r9, %rax adcq %rdi, %rbp adcq %rcx, %r13 adcq %r8, %r12 adcq $0, %rsi addq %rdx, %rbp adcq %r14, %r13 adcq %r11, %r12 adcq %r10, %rsi movq -16(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdi movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %rcx movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r11 movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r14 movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rax, %r9 movq %rdx, %rdi addq %r14, %rdi adcq %r11, %r10 adcq %rcx, %r8 adcq $0, %rbx addq %rbp, %r9 adcq %r13, %rdi adcq %r12, %r10 adcq %rsi, %r8 adcq $0, %rbx movq -8(%rsp), %rsi # 8-byte Reload imulq %r9, %rsi movq %rsi, %rax movq -56(%rsp), %r12 # 8-byte Reload mulq %r12 movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, %r13 movq %rsi, %rax movq -72(%rsp), %r14 # 8-byte Reload mulq %r14 movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %rbp movq %rsi, %rax movq -64(%rsp), %r11 # 8-byte Reload mulq %r11 movq %rdx, %r15 movq %rax, %rcx movq %rsi, %rax movq -80(%rsp), %rsi # 8-byte Reload mulq %rsi addq %r9, %rcx adcq %rdi, %rax adcq %r10, %rbp adcq %r8, %r13 adcq $0, %rbx addq %r15, %rax adcq %rdx, %rbp adcq -16(%rsp), %r13 # 8-byte Folded Reload adcq -8(%rsp), %rbx # 8-byte Folded Reload movq %rax, %rcx subq %r11, %rcx movq %rbp, %rdx sbbq %rsi, %rdx movq %r13, %rdi sbbq %r14, %rdi movq %rbx, %rsi sbbq %r12, %rsi cmovsq %rax, %rcx movq -88(%rsp), %rax # 8-byte Reload movq %rcx, (%rax) cmovsq %rbp, %rdx movq %rdx, 8(%rax) cmovsq %r13, %rdi movq %rdi, 16(%rax) cmovsq %rbx, %rsi movq %rsi, 24(%rax) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end54: .size mcl_fp_montNF4L, .Lfunc_end54-mcl_fp_montNF4L .globl mcl_fp_montRed4L .align 16, 0x90 .type mcl_fp_montRed4L,@function mcl_fp_montRed4L: # @mcl_fp_montRed4L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rdi, -56(%rsp) # 8-byte Spill movq -8(%rcx), %rax movq %rax, -40(%rsp) # 8-byte Spill movq (%rcx), %rdi movq %rdi, -48(%rsp) # 8-byte Spill movq (%rsi), %r12 movq %r12, %rbx imulq %rax, %rbx movq %rax, %r9 movq 24(%rcx), %rdx movq %rdx, -8(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rdx movq %rax, %r11 movq %rdx, %r8 movq 16(%rcx), %rbp movq %rbp, -32(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rbp movq %rbp, %r13 movq %rax, %r14 movq %rdx, %r10 movq 8(%rcx), %rcx movq %rcx, -24(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rcx movq %rcx, %rbp movq %rdx, %r15 movq %rax, %rcx movq %rbx, %rax mulq %rdi movq %rdx, %rbx addq %rcx, %rbx adcq %r14, %r15 adcq %r11, %r10 adcq $0, %r8 movq 56(%rsi), %rcx movq 48(%rsi), %rdx addq %r12, %rax movq 40(%rsi), %rax adcq 8(%rsi), %rbx adcq 16(%rsi), %r15 adcq 24(%rsi), %r10 adcq 32(%rsi), %r8 adcq $0, %rax movq %rax, -64(%rsp) # 8-byte Spill adcq $0, %rdx movq %rdx, %r12 adcq $0, %rcx movq %rcx, -16(%rsp) # 8-byte Spill sbbq %rdi, %rdi andl $1, %edi movq %rbx, %rsi imulq %r9, %rsi movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -72(%rsp) # 8-byte Spill movq %rsi, %rax mulq %r13 movq %rdx, %r14 movq %rax, %r9 movq %rsi, %rax mulq %rbp movq %rdx, %rcx movq %rax, %rbp movq %rsi, %rax movq -48(%rsp), %r13 # 8-byte Reload mulq %r13 movq %rdx, %rsi addq %rbp, %rsi adcq %r9, %rcx adcq -72(%rsp), %r14 # 8-byte Folded Reload adcq $0, %r11 addq %rbx, %rax adcq %r15, %rsi adcq %r10, %rcx adcq %r8, %r14 adcq -64(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r12 movq %r12, -64(%rsp) # 8-byte Spill movq -16(%rsp), %rbp # 8-byte Reload adcq $0, %rbp adcq $0, %rdi movq %rsi, %rbx imulq -40(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax movq -8(%rsp), %r12 # 8-byte Reload mulq %r12 movq %rdx, %r8 movq %rax, -16(%rsp) # 8-byte Spill movq %rbx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -72(%rsp) # 8-byte Spill movq %rbx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %r9 movq %rbx, %rax mulq %r13 movq %rdx, %rbx addq %r9, %rbx adcq -72(%rsp), %r15 # 8-byte Folded Reload adcq -16(%rsp), %r10 # 8-byte Folded Reload adcq $0, %r8 addq %rsi, %rax adcq %rcx, %rbx adcq %r14, %r15 adcq %r11, %r10 adcq -64(%rsp), %r8 # 8-byte Folded Reload adcq $0, %rbp movq %rbp, -16(%rsp) # 8-byte Spill adcq $0, %rdi movq -40(%rsp), %rcx # 8-byte Reload imulq %rbx, %rcx movq %rcx, %rax mulq %r12 movq %rdx, %r13 movq %rax, -40(%rsp) # 8-byte Spill movq %rcx, %rax movq -32(%rsp), %r14 # 8-byte Reload mulq %r14 movq %rdx, %r11 movq %rax, %r12 movq %rcx, %rax movq %rcx, %r9 movq -24(%rsp), %rsi # 8-byte Reload mulq %rsi movq %rdx, %rbp movq %rax, %rcx movq %r9, %rax movq -48(%rsp), %r9 # 8-byte Reload mulq %r9 addq %rcx, %rdx adcq %r12, %rbp adcq -40(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r13 addq %rbx, %rax adcq %r15, %rdx adcq %r10, %rbp adcq %r8, %r11 adcq -16(%rsp), %r13 # 8-byte Folded Reload adcq $0, %rdi movq %rdx, %rax subq %r9, %rax movq %rbp, %rcx sbbq %rsi, %rcx movq %r11, %rbx sbbq %r14, %rbx movq %r13, %rsi sbbq -8(%rsp), %rsi # 8-byte Folded Reload sbbq $0, %rdi andl $1, %edi cmovneq %r13, %rsi testb %dil, %dil cmovneq %rdx, %rax movq -56(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovneq %rbp, %rcx movq %rcx, 8(%rdx) cmovneq %r11, %rbx movq %rbx, 16(%rdx) movq %rsi, 24(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end55: .size mcl_fp_montRed4L, .Lfunc_end55-mcl_fp_montRed4L .globl mcl_fp_addPre4L .align 16, 0x90 .type mcl_fp_addPre4L,@function mcl_fp_addPre4L: # @mcl_fp_addPre4L # BB#0: movq 24(%rdx), %r8 movq 24(%rsi), %r9 movq 16(%rdx), %rax movq (%rdx), %rcx movq 8(%rdx), %rdx addq (%rsi), %rcx adcq 8(%rsi), %rdx adcq 16(%rsi), %rax movq %rcx, (%rdi) movq %rdx, 8(%rdi) movq %rax, 16(%rdi) adcq %r8, %r9 movq %r9, 24(%rdi) sbbq %rax, %rax andl $1, %eax retq .Lfunc_end56: .size mcl_fp_addPre4L, .Lfunc_end56-mcl_fp_addPre4L .globl mcl_fp_subPre4L .align 16, 0x90 .type mcl_fp_subPre4L,@function mcl_fp_subPre4L: # @mcl_fp_subPre4L # BB#0: movq 24(%rdx), %r8 movq 24(%rsi), %r9 movq 16(%rsi), %r10 movq (%rsi), %rcx movq 8(%rsi), %rsi xorl %eax, %eax subq (%rdx), %rcx sbbq 8(%rdx), %rsi sbbq 16(%rdx), %r10 movq %rcx, (%rdi) movq %rsi, 8(%rdi) movq %r10, 16(%rdi) sbbq %r8, %r9 movq %r9, 24(%rdi) sbbq $0, %rax andl $1, %eax retq .Lfunc_end57: .size mcl_fp_subPre4L, .Lfunc_end57-mcl_fp_subPre4L .globl mcl_fp_shr1_4L .align 16, 0x90 .type mcl_fp_shr1_4L,@function mcl_fp_shr1_4L: # @mcl_fp_shr1_4L # BB#0: movq 24(%rsi), %rax movq 16(%rsi), %rcx movq (%rsi), %rdx movq 8(%rsi), %rsi shrdq $1, %rsi, %rdx movq %rdx, (%rdi) shrdq $1, %rcx, %rsi movq %rsi, 8(%rdi) shrdq $1, %rax, %rcx movq %rcx, 16(%rdi) shrq %rax movq %rax, 24(%rdi) retq .Lfunc_end58: .size mcl_fp_shr1_4L, .Lfunc_end58-mcl_fp_shr1_4L .globl mcl_fp_add4L .align 16, 0x90 .type mcl_fp_add4L,@function mcl_fp_add4L: # @mcl_fp_add4L # BB#0: movq 24(%rdx), %r10 movq 24(%rsi), %r8 movq 16(%rdx), %r9 movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %r9 movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r9, 16(%rdi) adcq %r10, %r8 movq %r8, 24(%rdi) sbbq %rsi, %rsi andl $1, %esi subq (%rcx), %rax sbbq 8(%rcx), %rdx sbbq 16(%rcx), %r9 sbbq 24(%rcx), %r8 sbbq $0, %rsi testb $1, %sil jne .LBB59_2 # BB#1: # %nocarry movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r9, 16(%rdi) movq %r8, 24(%rdi) .LBB59_2: # %carry retq .Lfunc_end59: .size mcl_fp_add4L, .Lfunc_end59-mcl_fp_add4L .globl mcl_fp_addNF4L .align 16, 0x90 .type mcl_fp_addNF4L,@function mcl_fp_addNF4L: # @mcl_fp_addNF4L # BB#0: pushq %rbx movq 24(%rdx), %r8 movq 16(%rdx), %r9 movq (%rdx), %r11 movq 8(%rdx), %r10 addq (%rsi), %r11 adcq 8(%rsi), %r10 adcq 16(%rsi), %r9 adcq 24(%rsi), %r8 movq %r11, %rsi subq (%rcx), %rsi movq %r10, %rdx sbbq 8(%rcx), %rdx movq %r9, %rax sbbq 16(%rcx), %rax movq %r8, %rbx sbbq 24(%rcx), %rbx testq %rbx, %rbx cmovsq %r11, %rsi movq %rsi, (%rdi) cmovsq %r10, %rdx movq %rdx, 8(%rdi) cmovsq %r9, %rax movq %rax, 16(%rdi) cmovsq %r8, %rbx movq %rbx, 24(%rdi) popq %rbx retq .Lfunc_end60: .size mcl_fp_addNF4L, .Lfunc_end60-mcl_fp_addNF4L .globl mcl_fp_sub4L .align 16, 0x90 .type mcl_fp_sub4L,@function mcl_fp_sub4L: # @mcl_fp_sub4L # BB#0: movq 24(%rdx), %r10 movq 24(%rsi), %r8 movq 16(%rsi), %r9 movq (%rsi), %rax movq 8(%rsi), %r11 xorl %esi, %esi subq (%rdx), %rax sbbq 8(%rdx), %r11 sbbq 16(%rdx), %r9 movq %rax, (%rdi) movq %r11, 8(%rdi) movq %r9, 16(%rdi) sbbq %r10, %r8 movq %r8, 24(%rdi) sbbq $0, %rsi testb $1, %sil jne .LBB61_2 # BB#1: # %nocarry retq .LBB61_2: # %carry movq 24(%rcx), %r10 movq 8(%rcx), %rsi movq 16(%rcx), %rdx addq (%rcx), %rax movq %rax, (%rdi) adcq %r11, %rsi movq %rsi, 8(%rdi) adcq %r9, %rdx movq %rdx, 16(%rdi) adcq %r8, %r10 movq %r10, 24(%rdi) retq .Lfunc_end61: .size mcl_fp_sub4L, .Lfunc_end61-mcl_fp_sub4L .globl mcl_fp_subNF4L .align 16, 0x90 .type mcl_fp_subNF4L,@function mcl_fp_subNF4L: # @mcl_fp_subNF4L # BB#0: pushq %rbx movq 24(%rsi), %r11 movq 16(%rsi), %r8 movq (%rsi), %r9 movq 8(%rsi), %r10 subq (%rdx), %r9 sbbq 8(%rdx), %r10 sbbq 16(%rdx), %r8 sbbq 24(%rdx), %r11 movq %r11, %rdx sarq $63, %rdx movq 24(%rcx), %rsi andq %rdx, %rsi movq 16(%rcx), %rax andq %rdx, %rax movq 8(%rcx), %rbx andq %rdx, %rbx andq (%rcx), %rdx addq %r9, %rdx movq %rdx, (%rdi) adcq %r10, %rbx movq %rbx, 8(%rdi) adcq %r8, %rax movq %rax, 16(%rdi) adcq %r11, %rsi movq %rsi, 24(%rdi) popq %rbx retq .Lfunc_end62: .size mcl_fp_subNF4L, .Lfunc_end62-mcl_fp_subNF4L .globl mcl_fpDbl_add4L .align 16, 0x90 .type mcl_fpDbl_add4L,@function mcl_fpDbl_add4L: # @mcl_fpDbl_add4L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 56(%rdx), %r9 movq 56(%rsi), %r8 movq 48(%rdx), %r10 movq 48(%rsi), %r12 movq 40(%rdx), %r11 movq 32(%rdx), %r14 movq 24(%rdx), %r15 movq 16(%rdx), %rbx movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %rbx movq 40(%rsi), %r13 movq 24(%rsi), %rbp movq 32(%rsi), %rsi movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %rbx, 16(%rdi) adcq %r15, %rbp movq %rbp, 24(%rdi) adcq %r14, %rsi adcq %r11, %r13 adcq %r10, %r12 adcq %r9, %r8 sbbq %rax, %rax andl $1, %eax movq %rsi, %rdx subq (%rcx), %rdx movq %r13, %rbp sbbq 8(%rcx), %rbp movq %r12, %rbx sbbq 16(%rcx), %rbx movq %r8, %r9 sbbq 24(%rcx), %r9 sbbq $0, %rax andl $1, %eax cmovneq %rsi, %rdx movq %rdx, 32(%rdi) testb %al, %al cmovneq %r13, %rbp movq %rbp, 40(%rdi) cmovneq %r12, %rbx movq %rbx, 48(%rdi) cmovneq %r8, %r9 movq %r9, 56(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end63: .size mcl_fpDbl_add4L, .Lfunc_end63-mcl_fpDbl_add4L .globl mcl_fpDbl_sub4L .align 16, 0x90 .type mcl_fpDbl_sub4L,@function mcl_fpDbl_sub4L: # @mcl_fpDbl_sub4L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq 56(%rdx), %r9 movq 56(%rsi), %r8 movq 48(%rdx), %r10 movq 24(%rdx), %r11 movq (%rsi), %rbx xorl %eax, %eax subq (%rdx), %rbx movq %rbx, (%rdi) movq 8(%rsi), %rbx sbbq 8(%rdx), %rbx movq %rbx, 8(%rdi) movq 16(%rsi), %rbx sbbq 16(%rdx), %rbx movq %rbx, 16(%rdi) movq 24(%rsi), %rbx sbbq %r11, %rbx movq 40(%rdx), %r11 movq 32(%rdx), %rdx movq %rbx, 24(%rdi) movq 32(%rsi), %r12 sbbq %rdx, %r12 movq 48(%rsi), %r14 movq 40(%rsi), %r15 sbbq %r11, %r15 sbbq %r10, %r14 sbbq %r9, %r8 movl $0, %edx sbbq $0, %rdx andl $1, %edx movq (%rcx), %rsi cmoveq %rax, %rsi testb %dl, %dl movq 16(%rcx), %rdx cmoveq %rax, %rdx movq 24(%rcx), %rbx cmoveq %rax, %rbx cmovneq 8(%rcx), %rax addq %r12, %rsi movq %rsi, 32(%rdi) adcq %r15, %rax movq %rax, 40(%rdi) adcq %r14, %rdx movq %rdx, 48(%rdi) adcq %r8, %rbx movq %rbx, 56(%rdi) popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end64: .size mcl_fpDbl_sub4L, .Lfunc_end64-mcl_fpDbl_sub4L .globl mcl_fp_mulUnitPre5L .align 16, 0x90 .type mcl_fp_mulUnitPre5L,@function mcl_fp_mulUnitPre5L: # @mcl_fp_mulUnitPre5L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rcx, %rax mulq 32(%rsi) movq %rdx, %r8 movq %rax, %r9 movq %rcx, %rax mulq 24(%rsi) movq %rdx, %r10 movq %rax, %r11 movq %rcx, %rax mulq 16(%rsi) movq %rdx, %r15 movq %rax, %r14 movq %rcx, %rax mulq 8(%rsi) movq %rdx, %rbx movq %rax, %r12 movq %rcx, %rax mulq (%rsi) movq %rax, (%rdi) addq %r12, %rdx movq %rdx, 8(%rdi) adcq %r14, %rbx movq %rbx, 16(%rdi) adcq %r11, %r15 movq %r15, 24(%rdi) adcq %r9, %r10 movq %r10, 32(%rdi) adcq $0, %r8 movq %r8, 40(%rdi) popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end65: .size mcl_fp_mulUnitPre5L, .Lfunc_end65-mcl_fp_mulUnitPre5L .globl mcl_fpDbl_mulPre5L .align 16, 0x90 .type mcl_fpDbl_mulPre5L,@function mcl_fpDbl_mulPre5L: # @mcl_fpDbl_mulPre5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -16(%rsp) # 8-byte Spill movq %rsi, %r9 movq %rdi, -8(%rsp) # 8-byte Spill movq (%r9), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 8(%r9), %rbx movq %rbx, -48(%rsp) # 8-byte Spill movq (%rdx), %rbp movq %rdx, %r8 mulq %rbp movq %rdx, -32(%rsp) # 8-byte Spill movq 16(%r9), %r13 movq 24(%r9), %r14 movq 32(%r9), %r15 movq %rax, (%rdi) movq %r15, %rax mulq %rbp movq %rdx, %r10 movq %rax, -40(%rsp) # 8-byte Spill movq %r14, %rax mulq %rbp movq %rdx, %r12 movq %rax, %r11 movq %r13, %rax mulq %rbp movq %rdx, %rcx movq %rax, %rsi movq %rbx, %rax mulq %rbp movq %rdx, %rbp movq %rax, %rdi addq -32(%rsp), %rdi # 8-byte Folded Reload adcq %rsi, %rbp adcq %r11, %rcx adcq -40(%rsp), %r12 # 8-byte Folded Reload adcq $0, %r10 movq 8(%r8), %r11 movq %r15, %rax mulq %r11 movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %rsi movq %r14, %rax mulq %r11 movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %r15 movq %r13, %rax mulq %r11 movq %rdx, %r8 movq %rax, %r13 movq -48(%rsp), %rax # 8-byte Reload mulq %r11 movq %rdx, %r14 movq %rax, %rbx movq -24(%rsp), %rax # 8-byte Reload mulq %r11 addq %rdi, %rax movq -8(%rsp), %rdi # 8-byte Reload movq %rax, 8(%rdi) adcq %rbp, %rbx adcq %rcx, %r13 adcq %r12, %r15 adcq %r10, %rsi sbbq %rcx, %rcx andl $1, %ecx addq %rdx, %rbx adcq %r14, %r13 adcq %r8, %r15 adcq -40(%rsp), %rsi # 8-byte Folded Reload adcq -32(%rsp), %rcx # 8-byte Folded Reload movq 32(%r9), %rax movq %rax, -40(%rsp) # 8-byte Spill movq -16(%rsp), %rdi # 8-byte Reload movq 16(%rdi), %r12 mulq %r12 movq %rax, %r11 movq %rdx, -24(%rsp) # 8-byte Spill movq 24(%r9), %rax movq %rax, -72(%rsp) # 8-byte Spill mulq %r12 movq %rax, %r10 movq %rdx, -32(%rsp) # 8-byte Spill movq 16(%r9), %rax movq %rax, -80(%rsp) # 8-byte Spill mulq %r12 movq %rax, %r8 movq %rdx, -48(%rsp) # 8-byte Spill movq (%r9), %r14 movq 8(%r9), %rax movq %rax, -56(%rsp) # 8-byte Spill mulq %r12 movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %rbp movq %r14, %rax mulq %r12 movq %rdx, -88(%rsp) # 8-byte Spill addq %rbx, %rax movq -8(%rsp), %rbx # 8-byte Reload movq %rax, 16(%rbx) adcq %r13, %rbp adcq %r15, %r8 adcq %rsi, %r10 adcq %rcx, %r11 sbbq %rcx, %rcx movq 24(%rdi), %rsi movq -40(%rsp), %rax # 8-byte Reload mulq %rsi movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %r13 movq -56(%rsp), %rax # 8-byte Reload mulq %rsi movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %r12 movq %r14, %rax mulq %rsi movq %rdx, %r15 movq %rax, %rdi movq -72(%rsp), %rax # 8-byte Reload mulq %rsi movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %r14 movq -80(%rsp), %rax # 8-byte Reload mulq %rsi andl $1, %ecx addq -88(%rsp), %rbp # 8-byte Folded Reload adcq -64(%rsp), %r8 # 8-byte Folded Reload adcq -48(%rsp), %r10 # 8-byte Folded Reload adcq -32(%rsp), %r11 # 8-byte Folded Reload adcq -24(%rsp), %rcx # 8-byte Folded Reload addq %rdi, %rbp movq %rbp, 24(%rbx) adcq %r12, %r8 adcq %rax, %r10 adcq %r14, %r11 adcq %r13, %rcx sbbq %rsi, %rsi andl $1, %esi addq %r15, %r8 adcq -56(%rsp), %r10 # 8-byte Folded Reload adcq %rdx, %r11 adcq -72(%rsp), %rcx # 8-byte Folded Reload adcq -40(%rsp), %rsi # 8-byte Folded Reload movq -16(%rsp), %rax # 8-byte Reload movq 32(%rax), %rdi movq %rdi, %rax mulq 32(%r9) movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %r15 movq %rdi, %rax mulq 24(%r9) movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r13 movq %rdi, %rax mulq 16(%r9) movq %rdx, %r14 movq %rax, %rbx movq %rdi, %rax mulq 8(%r9) movq %rdx, %r12 movq %rax, %rbp movq %rdi, %rax mulq (%r9) addq %r8, %rax movq -8(%rsp), %rdi # 8-byte Reload movq %rax, 32(%rdi) adcq %r10, %rbp adcq %r11, %rbx adcq %rcx, %r13 adcq %rsi, %r15 sbbq %rax, %rax andl $1, %eax addq %rdx, %rbp movq %rbp, 40(%rdi) adcq %r12, %rbx movq %rbx, 48(%rdi) adcq %r14, %r13 movq %r13, 56(%rdi) adcq -24(%rsp), %r15 # 8-byte Folded Reload movq %r15, 64(%rdi) adcq -16(%rsp), %rax # 8-byte Folded Reload movq %rax, 72(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end66: .size mcl_fpDbl_mulPre5L, .Lfunc_end66-mcl_fpDbl_mulPre5L .globl mcl_fpDbl_sqrPre5L .align 16, 0x90 .type mcl_fpDbl_sqrPre5L,@function mcl_fpDbl_sqrPre5L: # @mcl_fpDbl_sqrPre5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdi, -8(%rsp) # 8-byte Spill movq 32(%rsi), %r11 movq (%rsi), %r13 movq 8(%rsi), %rbx movq %r11, %rax mulq %rbx movq %rax, -40(%rsp) # 8-byte Spill movq %rdx, -16(%rsp) # 8-byte Spill movq 24(%rsi), %rbp movq %rbp, %rax mulq %rbx movq %rax, -48(%rsp) # 8-byte Spill movq %rdx, -24(%rsp) # 8-byte Spill movq 16(%rsi), %rcx movq %rcx, %rax mulq %rbx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, -56(%rsp) # 8-byte Spill movq %r11, %rax mulq %r13 movq %rdx, %r8 movq %rax, -72(%rsp) # 8-byte Spill movq %rbp, %rax mulq %r13 movq %rdx, %r9 movq %rax, %rbp movq %rcx, %rax mulq %r13 movq %rdx, %r10 movq %rax, %r12 movq %rbx, %rax mulq %rbx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %r15 movq %rbx, %rax mulq %r13 movq %rdx, %rbx movq %rax, %r14 movq %r13, %rax mulq %r13 movq %rax, (%rdi) addq %r14, %rdx adcq %rbx, %r12 adcq %rbp, %r10 adcq -72(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r8 addq %r14, %rdx movq %rdx, 8(%rdi) adcq %r15, %r12 adcq -56(%rsp), %r10 # 8-byte Folded Reload adcq -48(%rsp), %r9 # 8-byte Folded Reload adcq -40(%rsp), %r8 # 8-byte Folded Reload sbbq %rdi, %rdi andl $1, %edi addq %rbx, %r12 adcq -64(%rsp), %r10 # 8-byte Folded Reload adcq -32(%rsp), %r9 # 8-byte Folded Reload adcq -24(%rsp), %r8 # 8-byte Folded Reload adcq -16(%rsp), %rdi # 8-byte Folded Reload movq %r11, %rax mulq %rcx movq %rax, %r11 movq %rdx, -16(%rsp) # 8-byte Spill movq 24(%rsi), %rbx movq %rbx, %rax mulq %rcx movq %rax, %r14 movq %rdx, -24(%rsp) # 8-byte Spill movq (%rsi), %rbp movq %rbp, -32(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -40(%rsp) # 8-byte Spill mulq %rcx movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %r15 movq %rbp, %rax mulq %rcx movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %rbp movq %rcx, %rax mulq %rcx movq %rax, %r13 addq %r12, %rbp movq -8(%rsp), %rax # 8-byte Reload movq %rbp, 16(%rax) adcq %r10, %r15 adcq %r9, %r13 adcq %r8, %r14 adcq %rdi, %r11 sbbq %r10, %r10 andl $1, %r10d addq -56(%rsp), %r15 # 8-byte Folded Reload adcq -48(%rsp), %r13 # 8-byte Folded Reload adcq %rdx, %r14 adcq -24(%rsp), %r11 # 8-byte Folded Reload adcq -16(%rsp), %r10 # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r8 movq -32(%rsp), %rax # 8-byte Reload mulq %rbx movq %rax, %rdi movq %rdx, -40(%rsp) # 8-byte Spill movq 16(%rsi), %rbp movq %rbp, -16(%rsp) # 8-byte Spill movq 32(%rsi), %rcx movq %rcx, %rax mulq %rbx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r9 movq %rbp, %rax mulq %rbx movq %rdx, %rbp movq %rax, %r12 movq %rbx, %rax mulq %rbx movq %rax, %rbx addq %r15, %rdi movq -8(%rsp), %r15 # 8-byte Reload movq %rdi, 24(%r15) adcq %r13, %r8 adcq %r14, %r12 adcq %r11, %rbx adcq %r10, %r9 sbbq %r10, %r10 andl $1, %r10d addq -40(%rsp), %r8 # 8-byte Folded Reload adcq -24(%rsp), %r12 # 8-byte Folded Reload adcq %rbp, %rbx adcq %rdx, %r9 adcq -32(%rsp), %r10 # 8-byte Folded Reload movq %rcx, %rax mulq 24(%rsi) movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %rbp movq %rcx, %rax mulq 8(%rsi) movq %rdx, %r14 movq %rax, %rdi movq %rcx, %rax mulq (%rsi) movq %rdx, %r13 movq %rax, %rsi movq %rcx, %rax mulq %rcx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r11 movq -16(%rsp), %rax # 8-byte Reload mulq %rcx addq %r8, %rsi movq %rsi, 32(%r15) adcq %r12, %rdi adcq %rbx, %rax adcq %r9, %rbp adcq %r10, %r11 sbbq %rcx, %rcx andl $1, %ecx addq %r13, %rdi movq %rdi, 40(%r15) adcq %r14, %rax movq %rax, 48(%r15) adcq %rdx, %rbp movq %rbp, 56(%r15) adcq -24(%rsp), %r11 # 8-byte Folded Reload movq %r11, 64(%r15) adcq -32(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 72(%r15) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end67: .size mcl_fpDbl_sqrPre5L, .Lfunc_end67-mcl_fpDbl_sqrPre5L .globl mcl_fp_mont5L .align 16, 0x90 .type mcl_fp_mont5L,@function mcl_fp_mont5L: # @mcl_fp_mont5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx pushq %rax movq %rdx, -48(%rsp) # 8-byte Spill movq %rdi, -104(%rsp) # 8-byte Spill movq 32(%rsi), %rax movq %rax, -56(%rsp) # 8-byte Spill movq (%rdx), %rdi mulq %rdi movq %rax, %r8 movq %rdx, %r14 movq 24(%rsi), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rdi movq %rax, %r9 movq %rdx, %r12 movq 16(%rsi), %rax movq %rax, -72(%rsp) # 8-byte Spill mulq %rdi movq %rax, %r10 movq %rdx, %rbp movq (%rsi), %rbx movq %rbx, -80(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -88(%rsp) # 8-byte Spill mulq %rdi movq %rdx, %r11 movq %rax, %rsi movq %rbx, %rax mulq %rdi movq %rax, -128(%rsp) # 8-byte Spill movq %rdx, %r15 addq %rsi, %r15 adcq %r10, %r11 adcq %r9, %rbp movq %rbp, -96(%rsp) # 8-byte Spill adcq %r8, %r12 movq %r12, -112(%rsp) # 8-byte Spill adcq $0, %r14 movq %r14, -120(%rsp) # 8-byte Spill movq -8(%rcx), %rdx movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %rbp imulq %rdx, %rbp movq (%rcx), %r9 movq %r9, -32(%rsp) # 8-byte Spill movq 32(%rcx), %rdx movq %rdx, (%rsp) # 8-byte Spill movq 24(%rcx), %rsi movq %rsi, -8(%rsp) # 8-byte Spill movq 16(%rcx), %rbx movq %rbx, -16(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -24(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rdx, %r14 movq %rax, %r13 movq %rbp, %rax mulq %rsi movq %rdx, %rdi movq %rax, %r10 movq %rbp, %rax mulq %rbx movq %rdx, %rbx movq %rax, %r8 movq %rbp, %rax mulq %rcx movq %rdx, %rsi movq %rax, %r12 movq %rbp, %rax mulq %r9 movq %rdx, %rbp addq %r12, %rbp adcq %r8, %rsi adcq %r10, %rbx adcq %r13, %rdi adcq $0, %r14 addq -128(%rsp), %rax # 8-byte Folded Reload adcq %r15, %rbp adcq %r11, %rsi adcq -96(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %rdi # 8-byte Folded Reload adcq -120(%rsp), %r14 # 8-byte Folded Reload sbbq %r9, %r9 andl $1, %r9d movq -48(%rsp), %rax # 8-byte Reload movq 8(%rax), %rcx movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -96(%rsp) # 8-byte Spill movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r10 movq %rcx, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r15 movq %rdx, %rcx addq %r10, %rcx adcq -120(%rsp), %r8 # 8-byte Folded Reload adcq -112(%rsp), %r12 # 8-byte Folded Reload adcq -96(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r13 addq %rbp, %r15 adcq %rsi, %rcx adcq %rbx, %r8 adcq %rdi, %r12 adcq %r14, %r11 adcq %r9, %r13 sbbq %rax, %rax andl $1, %eax movq %rax, -96(%rsp) # 8-byte Spill movq %r15, %rsi imulq -40(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -128(%rsp) # 8-byte Spill movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rdi movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rbx addq %rdi, %rbx adcq -128(%rsp), %r10 # 8-byte Folded Reload adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq -112(%rsp), %rbp # 8-byte Folded Reload adcq $0, %r14 addq %r15, %rax adcq %rcx, %rbx adcq %r8, %r10 adcq %r12, %r9 adcq %r11, %rbp adcq %r13, %r14 adcq $0, -96(%rsp) # 8-byte Folded Spill movq -48(%rsp), %rax # 8-byte Reload movq 16(%rax), %rcx movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, -128(%rsp) # 8-byte Spill movq %rcx, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r8 movq %rcx, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r12 movq %rdx, %r15 addq %r8, %r15 adcq -128(%rsp), %rdi # 8-byte Folded Reload adcq -120(%rsp), %rsi # 8-byte Folded Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r13 addq %rbx, %r12 adcq %r10, %r15 adcq %r9, %rdi adcq %rbp, %rsi adcq %r14, %r11 adcq -96(%rsp), %r13 # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax movq %rax, -96(%rsp) # 8-byte Spill movq %r12, %rbp imulq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r10 movq %rbp, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r14 movq %rbp, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rbp addq %r14, %rbp adcq %r10, %rbx adcq -120(%rsp), %rcx # 8-byte Folded Reload adcq -112(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r8 addq %r12, %rax adcq %r15, %rbp adcq %rdi, %rbx adcq %rsi, %rcx adcq %r11, %r9 adcq %r13, %r8 adcq $0, -96(%rsp) # 8-byte Folded Spill movq -48(%rsp), %rax # 8-byte Reload movq 24(%rax), %rsi movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %r15 movq %rsi, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r12 movq %rsi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r14 movq %rdx, %rsi addq %r12, %rsi adcq %r15, %rdi adcq -120(%rsp), %r11 # 8-byte Folded Reload adcq -112(%rsp), %r10 # 8-byte Folded Reload adcq $0, %r13 addq %rbp, %r14 adcq %rbx, %rsi adcq %rcx, %rdi adcq %r9, %r11 adcq %r8, %r10 adcq -96(%rsp), %r13 # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax movq %rax, -96(%rsp) # 8-byte Spill movq %r14, %rbp imulq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r8 movq %rbp, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r12 movq %rbp, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rbp addq %r12, %rbp adcq %r8, %rbx adcq -120(%rsp), %rcx # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload adcq $0, %r9 addq %r14, %rax adcq %rsi, %rbp adcq %rdi, %rbx adcq %r11, %rcx adcq %r10, %r15 adcq %r13, %r9 movq -96(%rsp), %r14 # 8-byte Reload adcq $0, %r14 movq -48(%rsp), %rax # 8-byte Reload movq 32(%rax), %rsi movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -56(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -64(%rsp) # 8-byte Spill movq %rsi, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -72(%rsp) # 8-byte Spill movq %rsi, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %rdi movq %rsi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r10 movq %rdx, %r8 addq %rdi, %r8 adcq -72(%rsp), %r12 # 8-byte Folded Reload adcq -64(%rsp), %r11 # 8-byte Folded Reload adcq -56(%rsp), %r13 # 8-byte Folded Reload movq -48(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rbp, %r10 adcq %rbx, %r8 adcq %rcx, %r12 adcq %r15, %r11 adcq %r9, %r13 adcq %r14, %rax movq %rax, -48(%rsp) # 8-byte Spill sbbq %rcx, %rcx movq -40(%rsp), %rsi # 8-byte Reload imulq %r10, %rsi movq %rsi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -40(%rsp) # 8-byte Spill movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, -56(%rsp) # 8-byte Spill movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r15 movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r9 movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload addq %r9, %rdx adcq %r15, %rdi adcq -56(%rsp), %rbp # 8-byte Folded Reload adcq -40(%rsp), %rbx # 8-byte Folded Reload adcq $0, %r14 andl $1, %ecx addq %r10, %rax adcq %r8, %rdx adcq %r12, %rdi adcq %r11, %rbp adcq %r13, %rbx adcq -48(%rsp), %r14 # 8-byte Folded Reload adcq $0, %rcx movq %rdx, %rax subq -32(%rsp), %rax # 8-byte Folded Reload movq %rdi, %r8 sbbq -24(%rsp), %r8 # 8-byte Folded Reload movq %rbp, %r9 sbbq -16(%rsp), %r9 # 8-byte Folded Reload movq %rbx, %r10 sbbq -8(%rsp), %r10 # 8-byte Folded Reload movq %r14, %r11 sbbq (%rsp), %r11 # 8-byte Folded Reload sbbq $0, %rcx andl $1, %ecx cmovneq %rbx, %r10 testb %cl, %cl cmovneq %rdx, %rax movq -104(%rsp), %rcx # 8-byte Reload movq %rax, (%rcx) cmovneq %rdi, %r8 movq %r8, 8(%rcx) cmovneq %rbp, %r9 movq %r9, 16(%rcx) movq %r10, 24(%rcx) cmovneq %r14, %r11 movq %r11, 32(%rcx) addq $8, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end68: .size mcl_fp_mont5L, .Lfunc_end68-mcl_fp_mont5L .globl mcl_fp_montNF5L .align 16, 0x90 .type mcl_fp_montNF5L,@function mcl_fp_montNF5L: # @mcl_fp_montNF5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -40(%rsp) # 8-byte Spill movq %rdi, -104(%rsp) # 8-byte Spill movq 32(%rsi), %rax movq %rax, -48(%rsp) # 8-byte Spill movq (%rdx), %rbx mulq %rbx movq %rax, %r15 movq %rdx, %r10 movq 24(%rsi), %rax movq %rax, -56(%rsp) # 8-byte Spill mulq %rbx movq %rax, %r13 movq %rdx, %r14 movq 16(%rsi), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rbx movq %rax, %r8 movq %rdx, %r9 movq (%rsi), %rbp movq %rbp, -80(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -88(%rsp) # 8-byte Spill mulq %rbx movq %rdx, %r11 movq %rax, %rdi movq %rbp, %rax mulq %rbx movq %rax, -112(%rsp) # 8-byte Spill movq %rdx, %r12 addq %rdi, %r12 adcq %r8, %r11 adcq %r13, %r9 adcq %r15, %r14 adcq $0, %r10 movq -8(%rcx), %rdx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %rsi imulq %rdx, %rsi movq (%rcx), %r8 movq %r8, -96(%rsp) # 8-byte Spill movq 32(%rcx), %rdx movq %rdx, -8(%rsp) # 8-byte Spill movq 24(%rcx), %rdi movq %rdi, -16(%rsp) # 8-byte Spill movq 16(%rcx), %rbx movq %rbx, -24(%rsp) # 8-byte Spill movq 8(%rcx), %rbp movq %rbp, -72(%rsp) # 8-byte Spill movq %rsi, %rax mulq %rdx movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %rcx movq %rsi, %rax mulq %rdi movq %rdx, -128(%rsp) # 8-byte Spill movq %rax, %rdi movq %rsi, %rax mulq %rbx movq %rdx, %r15 movq %rax, %rbx movq %rsi, %rax mulq %rbp movq %rdx, %r13 movq %rax, %rbp movq %rsi, %rax mulq %r8 addq -112(%rsp), %rax # 8-byte Folded Reload adcq %r12, %rbp adcq %r11, %rbx adcq %r9, %rdi adcq %r14, %rcx adcq $0, %r10 addq %rdx, %rbp adcq %r13, %rbx adcq %r15, %rdi adcq -128(%rsp), %rcx # 8-byte Folded Reload adcq -120(%rsp), %r10 # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload movq 8(%rax), %rsi movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %r8 movq %rsi, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %r14 movq %rsi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %rsi movq %rdx, %r12 addq %r14, %r12 adcq %r8, %r11 adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload adcq $0, %r13 addq %rbp, %rsi adcq %rbx, %r12 adcq %rdi, %r11 adcq %rcx, %r9 adcq %r10, %r15 adcq $0, %r13 movq %rsi, %rdi imulq -32(%rsp), %rdi # 8-byte Folded Reload movq %rdi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %rbp movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r14 movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r8 movq %rdi, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r10 movq %rdi, %rax mulq -96(%rsp) # 8-byte Folded Reload addq %rsi, %rax adcq %r12, %r10 adcq %r11, %r8 adcq %r9, %r14 adcq %r15, %rbp adcq $0, %r13 addq %rdx, %r10 adcq %rbx, %r8 adcq %rcx, %r14 adcq -120(%rsp), %rbp # 8-byte Folded Reload adcq -112(%rsp), %r13 # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload movq 16(%rax), %rsi movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %rbx movq %rsi, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r12 movq %rsi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r11 movq %rdx, %rsi addq %r12, %rsi adcq %rbx, %rcx adcq -120(%rsp), %rdi # 8-byte Folded Reload adcq -112(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r15 addq %r10, %r11 adcq %r8, %rsi adcq %r14, %rcx adcq %rbp, %rdi adcq %r13, %r9 adcq $0, %r15 movq %r11, %rbx imulq -32(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r13 movq %rbx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r8 movq %rbx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %r10 movq %rbx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %rbp movq %rbx, %rax mulq -96(%rsp) # 8-byte Folded Reload addq %r11, %rax adcq %rsi, %rbp adcq %rcx, %r10 adcq %rdi, %r8 adcq %r9, %r13 adcq $0, %r15 addq %rdx, %rbp adcq %r12, %r10 adcq %r14, %r8 adcq -120(%rsp), %r13 # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload movq 24(%rax), %rsi movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %rbx movq %rsi, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r12 movq %rsi, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r14 movq %rdx, %rsi addq %r12, %rsi adcq %rbx, %rcx adcq -120(%rsp), %rdi # 8-byte Folded Reload adcq -112(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r11 addq %rbp, %r14 adcq %r10, %rsi adcq %r8, %rcx adcq %r13, %rdi adcq %r15, %r9 adcq $0, %r11 movq %r14, %rbx imulq -32(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r13 movq %rbx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r8 movq %rbx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %r10 movq %rbx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %rbp movq %rbx, %rax mulq -96(%rsp) # 8-byte Folded Reload addq %r14, %rax adcq %rsi, %rbp adcq %rcx, %r10 adcq %rdi, %r8 adcq %r9, %r13 adcq $0, %r11 addq %rdx, %rbp adcq %r12, %r10 adcq %r15, %r8 adcq -120(%rsp), %r13 # 8-byte Folded Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload movq 32(%rax), %rcx movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, -40(%rsp) # 8-byte Spill movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -48(%rsp) # 8-byte Spill movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -56(%rsp) # 8-byte Spill movq %rcx, %rax mulq -88(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %rsi movq %rcx, %rax mulq -80(%rsp) # 8-byte Folded Reload movq %rax, %r12 movq %rdx, %rdi addq %rsi, %rdi adcq -56(%rsp), %r15 # 8-byte Folded Reload adcq -48(%rsp), %r14 # 8-byte Folded Reload adcq -40(%rsp), %r9 # 8-byte Folded Reload adcq $0, %rbx addq %rbp, %r12 adcq %r10, %rdi adcq %r8, %r15 adcq %r13, %r14 adcq %r11, %r9 adcq $0, %rbx movq -32(%rsp), %r8 # 8-byte Reload imulq %r12, %r8 movq %r8, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %rcx movq %r8, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %rbp movq %r8, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %rsi movq %r8, %rax movq %r8, %r13 movq -96(%rsp), %r10 # 8-byte Reload mulq %r10 movq %rdx, %r11 movq %rax, %r8 movq %r13, %rax movq -72(%rsp), %r13 # 8-byte Reload mulq %r13 addq %r12, %r8 adcq %rdi, %rax adcq %r15, %rsi adcq %r14, %rbp adcq %r9, %rcx adcq $0, %rbx addq %r11, %rax adcq %rdx, %rsi adcq -48(%rsp), %rbp # 8-byte Folded Reload adcq -40(%rsp), %rcx # 8-byte Folded Reload adcq -32(%rsp), %rbx # 8-byte Folded Reload movq %rax, %r11 subq %r10, %r11 movq %rsi, %r10 sbbq %r13, %r10 movq %rbp, %r8 sbbq -24(%rsp), %r8 # 8-byte Folded Reload movq %rcx, %r9 sbbq -16(%rsp), %r9 # 8-byte Folded Reload movq %rbx, %rdx sbbq -8(%rsp), %rdx # 8-byte Folded Reload movq %rdx, %rdi sarq $63, %rdi cmovsq %rax, %r11 movq -104(%rsp), %rax # 8-byte Reload movq %r11, (%rax) cmovsq %rsi, %r10 movq %r10, 8(%rax) cmovsq %rbp, %r8 movq %r8, 16(%rax) cmovsq %rcx, %r9 movq %r9, 24(%rax) cmovsq %rbx, %rdx movq %rdx, 32(%rax) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end69: .size mcl_fp_montNF5L, .Lfunc_end69-mcl_fp_montNF5L .globl mcl_fp_montRed5L .align 16, 0x90 .type mcl_fp_montRed5L,@function mcl_fp_montRed5L: # @mcl_fp_montRed5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rdi, -80(%rsp) # 8-byte Spill movq -8(%rcx), %rax movq %rax, -64(%rsp) # 8-byte Spill movq (%rcx), %rdi movq %rdi, -24(%rsp) # 8-byte Spill movq (%rsi), %r9 movq %r9, %rbp imulq %rax, %rbp movq 32(%rcx), %rdx movq %rdx, -48(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rax, %r10 movq %rdx, %r13 movq 24(%rcx), %rdx movq %rdx, -8(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rax, %r14 movq %rdx, %r11 movq 16(%rcx), %rdx movq %rdx, -16(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -32(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rdx, %r15 movq %rax, %r12 movq %rbp, %rax mulq %rcx movq %rdx, %r8 movq %rax, %rbx movq %rbp, %rax mulq %rdi movq %rdx, %rcx addq %rbx, %rcx adcq %r12, %r8 adcq %r14, %r15 adcq %r10, %r11 adcq $0, %r13 addq %r9, %rax movq 72(%rsi), %rax movq 64(%rsi), %rdx adcq 8(%rsi), %rcx adcq 16(%rsi), %r8 adcq 24(%rsi), %r15 adcq 32(%rsi), %r11 adcq 40(%rsi), %r13 movq %r13, -88(%rsp) # 8-byte Spill movq 56(%rsi), %rdi movq 48(%rsi), %rsi adcq $0, %rsi movq %rsi, -96(%rsp) # 8-byte Spill adcq $0, %rdi movq %rdi, -72(%rsp) # 8-byte Spill adcq $0, %rdx movq %rdx, -56(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, -40(%rsp) # 8-byte Spill sbbq %rdi, %rdi andl $1, %edi movq %rcx, %rsi movq -64(%rsp), %r9 # 8-byte Reload imulq %r9, %rsi movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %rbx movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %rbp movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rsi addq %rbp, %rsi adcq %rbx, %r13 adcq -112(%rsp), %r12 # 8-byte Folded Reload adcq -104(%rsp), %r14 # 8-byte Folded Reload adcq $0, %r10 addq %rcx, %rax adcq %r8, %rsi adcq %r15, %r13 adcq %r11, %r12 adcq -88(%rsp), %r14 # 8-byte Folded Reload adcq -96(%rsp), %r10 # 8-byte Folded Reload adcq $0, -72(%rsp) # 8-byte Folded Spill adcq $0, -56(%rsp) # 8-byte Folded Spill adcq $0, -40(%rsp) # 8-byte Folded Spill adcq $0, %rdi movq %rsi, %rcx imulq %r9, %rcx movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -88(%rsp) # 8-byte Spill movq %rcx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -96(%rsp) # 8-byte Spill movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -104(%rsp) # 8-byte Spill movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r8 movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rbp addq %r8, %rbp adcq -104(%rsp), %rbx # 8-byte Folded Reload adcq -96(%rsp), %r15 # 8-byte Folded Reload adcq -88(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r9 addq %rsi, %rax adcq %r13, %rbp adcq %r12, %rbx adcq %r14, %r15 adcq %r10, %r11 adcq -72(%rsp), %r9 # 8-byte Folded Reload adcq $0, -56(%rsp) # 8-byte Folded Spill adcq $0, -40(%rsp) # 8-byte Folded Spill adcq $0, %rdi movq %rbp, %rcx imulq -64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, %rax movq -48(%rsp), %rsi # 8-byte Reload mulq %rsi movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rcx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -96(%rsp) # 8-byte Spill movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %r10 movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r8 movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx addq %r8, %rcx adcq %r10, %r13 adcq -96(%rsp), %r12 # 8-byte Folded Reload adcq -88(%rsp), %r14 # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %rbp, %rax adcq %rbx, %rcx adcq %r15, %r13 adcq %r11, %r12 adcq %r9, %r14 adcq -56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill adcq $0, -40(%rsp) # 8-byte Folded Spill adcq $0, %rdi movq -64(%rsp), %rbx # 8-byte Reload imulq %rcx, %rbx movq %rbx, %rax mulq %rsi movq %rdx, %rsi movq %rax, -56(%rsp) # 8-byte Spill movq %rbx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -64(%rsp) # 8-byte Spill movq %rbx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r15 movq %rbx, %rax movq %rbx, %r10 movq -32(%rsp), %r11 # 8-byte Reload mulq %r11 movq %rdx, %rbx movq %rax, %r8 movq %r10, %rax movq -24(%rsp), %r10 # 8-byte Reload mulq %r10 addq %r8, %rdx adcq %r15, %rbx adcq -64(%rsp), %rbp # 8-byte Folded Reload adcq -56(%rsp), %r9 # 8-byte Folded Reload adcq $0, %rsi addq %rcx, %rax adcq %r13, %rdx adcq %r12, %rbx adcq %r14, %rbp adcq -72(%rsp), %r9 # 8-byte Folded Reload adcq -40(%rsp), %rsi # 8-byte Folded Reload adcq $0, %rdi movq %rdx, %rax subq %r10, %rax movq %rbx, %rcx sbbq %r11, %rcx movq %rbp, %r8 sbbq -16(%rsp), %r8 # 8-byte Folded Reload movq %r9, %r10 sbbq -8(%rsp), %r10 # 8-byte Folded Reload movq %rsi, %r11 sbbq -48(%rsp), %r11 # 8-byte Folded Reload sbbq $0, %rdi andl $1, %edi cmovneq %rsi, %r11 testb %dil, %dil cmovneq %rdx, %rax movq -80(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovneq %rbx, %rcx movq %rcx, 8(%rdx) cmovneq %rbp, %r8 movq %r8, 16(%rdx) cmovneq %r9, %r10 movq %r10, 24(%rdx) movq %r11, 32(%rdx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end70: .size mcl_fp_montRed5L, .Lfunc_end70-mcl_fp_montRed5L .globl mcl_fp_addPre5L .align 16, 0x90 .type mcl_fp_addPre5L,@function mcl_fp_addPre5L: # @mcl_fp_addPre5L # BB#0: movq 32(%rdx), %r8 movq 24(%rdx), %r9 movq 24(%rsi), %r11 movq 32(%rsi), %r10 movq 16(%rdx), %rcx movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %rcx movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %rcx, 16(%rdi) adcq %r9, %r11 movq %r11, 24(%rdi) adcq %r8, %r10 movq %r10, 32(%rdi) sbbq %rax, %rax andl $1, %eax retq .Lfunc_end71: .size mcl_fp_addPre5L, .Lfunc_end71-mcl_fp_addPre5L .globl mcl_fp_subPre5L .align 16, 0x90 .type mcl_fp_subPre5L,@function mcl_fp_subPre5L: # @mcl_fp_subPre5L # BB#0: pushq %rbx movq 32(%rsi), %r10 movq 24(%rdx), %r8 movq 32(%rdx), %r9 movq 24(%rsi), %r11 movq 16(%rsi), %rcx movq (%rsi), %rbx movq 8(%rsi), %rsi xorl %eax, %eax subq (%rdx), %rbx sbbq 8(%rdx), %rsi sbbq 16(%rdx), %rcx movq %rbx, (%rdi) movq %rsi, 8(%rdi) movq %rcx, 16(%rdi) sbbq %r8, %r11 movq %r11, 24(%rdi) sbbq %r9, %r10 movq %r10, 32(%rdi) sbbq $0, %rax andl $1, %eax popq %rbx retq .Lfunc_end72: .size mcl_fp_subPre5L, .Lfunc_end72-mcl_fp_subPre5L .globl mcl_fp_shr1_5L .align 16, 0x90 .type mcl_fp_shr1_5L,@function mcl_fp_shr1_5L: # @mcl_fp_shr1_5L # BB#0: movq 32(%rsi), %r8 movq 24(%rsi), %rcx movq 16(%rsi), %rdx movq (%rsi), %rax movq 8(%rsi), %rsi shrdq $1, %rsi, %rax movq %rax, (%rdi) shrdq $1, %rdx, %rsi movq %rsi, 8(%rdi) shrdq $1, %rcx, %rdx movq %rdx, 16(%rdi) shrdq $1, %r8, %rcx movq %rcx, 24(%rdi) shrq %r8 movq %r8, 32(%rdi) retq .Lfunc_end73: .size mcl_fp_shr1_5L, .Lfunc_end73-mcl_fp_shr1_5L .globl mcl_fp_add5L .align 16, 0x90 .type mcl_fp_add5L,@function mcl_fp_add5L: # @mcl_fp_add5L # BB#0: pushq %rbx movq 32(%rdx), %r11 movq 24(%rdx), %rbx movq 24(%rsi), %r9 movq 32(%rsi), %r8 movq 16(%rdx), %r10 movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %r10 movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r10, 16(%rdi) adcq %rbx, %r9 movq %r9, 24(%rdi) adcq %r11, %r8 movq %r8, 32(%rdi) sbbq %rsi, %rsi andl $1, %esi subq (%rcx), %rax sbbq 8(%rcx), %rdx sbbq 16(%rcx), %r10 sbbq 24(%rcx), %r9 sbbq 32(%rcx), %r8 sbbq $0, %rsi testb $1, %sil jne .LBB74_2 # BB#1: # %nocarry movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r10, 16(%rdi) movq %r9, 24(%rdi) movq %r8, 32(%rdi) .LBB74_2: # %carry popq %rbx retq .Lfunc_end74: .size mcl_fp_add5L, .Lfunc_end74-mcl_fp_add5L .globl mcl_fp_addNF5L .align 16, 0x90 .type mcl_fp_addNF5L,@function mcl_fp_addNF5L: # @mcl_fp_addNF5L # BB#0: pushq %r15 pushq %r14 pushq %rbx movq 32(%rdx), %r8 movq 24(%rdx), %r9 movq 16(%rdx), %r10 movq (%rdx), %r14 movq 8(%rdx), %r11 addq (%rsi), %r14 adcq 8(%rsi), %r11 adcq 16(%rsi), %r10 adcq 24(%rsi), %r9 adcq 32(%rsi), %r8 movq %r14, %rsi subq (%rcx), %rsi movq %r11, %rdx sbbq 8(%rcx), %rdx movq %r10, %rbx sbbq 16(%rcx), %rbx movq %r9, %r15 sbbq 24(%rcx), %r15 movq %r8, %rax sbbq 32(%rcx), %rax movq %rax, %rcx sarq $63, %rcx cmovsq %r14, %rsi movq %rsi, (%rdi) cmovsq %r11, %rdx movq %rdx, 8(%rdi) cmovsq %r10, %rbx movq %rbx, 16(%rdi) cmovsq %r9, %r15 movq %r15, 24(%rdi) cmovsq %r8, %rax movq %rax, 32(%rdi) popq %rbx popq %r14 popq %r15 retq .Lfunc_end75: .size mcl_fp_addNF5L, .Lfunc_end75-mcl_fp_addNF5L .globl mcl_fp_sub5L .align 16, 0x90 .type mcl_fp_sub5L,@function mcl_fp_sub5L: # @mcl_fp_sub5L # BB#0: pushq %r14 pushq %rbx movq 32(%rsi), %r8 movq 24(%rdx), %r11 movq 32(%rdx), %r14 movq 24(%rsi), %r9 movq 16(%rsi), %r10 movq (%rsi), %rax movq 8(%rsi), %rsi xorl %ebx, %ebx subq (%rdx), %rax sbbq 8(%rdx), %rsi sbbq 16(%rdx), %r10 movq %rax, (%rdi) movq %rsi, 8(%rdi) movq %r10, 16(%rdi) sbbq %r11, %r9 movq %r9, 24(%rdi) sbbq %r14, %r8 movq %r8, 32(%rdi) sbbq $0, %rbx testb $1, %bl je .LBB76_2 # BB#1: # %carry movq 32(%rcx), %r11 movq 24(%rcx), %r14 movq 8(%rcx), %rdx movq 16(%rcx), %rbx addq (%rcx), %rax movq %rax, (%rdi) adcq %rsi, %rdx movq %rdx, 8(%rdi) adcq %r10, %rbx movq %rbx, 16(%rdi) adcq %r9, %r14 movq %r14, 24(%rdi) adcq %r8, %r11 movq %r11, 32(%rdi) .LBB76_2: # %nocarry popq %rbx popq %r14 retq .Lfunc_end76: .size mcl_fp_sub5L, .Lfunc_end76-mcl_fp_sub5L .globl mcl_fp_subNF5L .align 16, 0x90 .type mcl_fp_subNF5L,@function mcl_fp_subNF5L: # @mcl_fp_subNF5L # BB#0: pushq %r15 pushq %r14 pushq %rbx movq 32(%rsi), %r14 movq 24(%rsi), %r8 movq 16(%rsi), %r9 movq (%rsi), %r10 movq 8(%rsi), %r11 subq (%rdx), %r10 sbbq 8(%rdx), %r11 sbbq 16(%rdx), %r9 sbbq 24(%rdx), %r8 sbbq 32(%rdx), %r14 movq %r14, %rdx sarq $63, %rdx movq %rdx, %rsi shldq $1, %r14, %rsi movq 8(%rcx), %rbx andq %rsi, %rbx andq (%rcx), %rsi movq 32(%rcx), %r15 andq %rdx, %r15 movq 24(%rcx), %rax andq %rdx, %rax rolq %rdx andq 16(%rcx), %rdx addq %r10, %rsi movq %rsi, (%rdi) adcq %r11, %rbx movq %rbx, 8(%rdi) adcq %r9, %rdx movq %rdx, 16(%rdi) adcq %r8, %rax movq %rax, 24(%rdi) adcq %r14, %r15 movq %r15, 32(%rdi) popq %rbx popq %r14 popq %r15 retq .Lfunc_end77: .size mcl_fp_subNF5L, .Lfunc_end77-mcl_fp_subNF5L .globl mcl_fpDbl_add5L .align 16, 0x90 .type mcl_fpDbl_add5L,@function mcl_fpDbl_add5L: # @mcl_fpDbl_add5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 72(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 64(%rdx), %r11 movq 56(%rdx), %r14 movq 48(%rdx), %r15 movq 24(%rsi), %rbp movq 32(%rsi), %r13 movq 16(%rdx), %r12 movq (%rdx), %rbx movq 8(%rdx), %rax addq (%rsi), %rbx adcq 8(%rsi), %rax adcq 16(%rsi), %r12 adcq 24(%rdx), %rbp adcq 32(%rdx), %r13 movq 40(%rdx), %r9 movq %rbx, (%rdi) movq 72(%rsi), %r8 movq %rax, 8(%rdi) movq 64(%rsi), %r10 movq %r12, 16(%rdi) movq 56(%rsi), %r12 movq %rbp, 24(%rdi) movq 48(%rsi), %rbp movq 40(%rsi), %rbx movq %r13, 32(%rdi) adcq %r9, %rbx adcq %r15, %rbp adcq %r14, %r12 adcq %r11, %r10 adcq -8(%rsp), %r8 # 8-byte Folded Reload sbbq %rsi, %rsi andl $1, %esi movq %rbx, %rax subq (%rcx), %rax movq %rbp, %rdx sbbq 8(%rcx), %rdx movq %r12, %r9 sbbq 16(%rcx), %r9 movq %r10, %r11 sbbq 24(%rcx), %r11 movq %r8, %r14 sbbq 32(%rcx), %r14 sbbq $0, %rsi andl $1, %esi cmovneq %rbx, %rax movq %rax, 40(%rdi) testb %sil, %sil cmovneq %rbp, %rdx movq %rdx, 48(%rdi) cmovneq %r12, %r9 movq %r9, 56(%rdi) cmovneq %r10, %r11 movq %r11, 64(%rdi) cmovneq %r8, %r14 movq %r14, 72(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end78: .size mcl_fpDbl_add5L, .Lfunc_end78-mcl_fpDbl_add5L .globl mcl_fpDbl_sub5L .align 16, 0x90 .type mcl_fpDbl_sub5L,@function mcl_fpDbl_sub5L: # @mcl_fpDbl_sub5L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 72(%rdx), %r9 movq 64(%rdx), %r10 movq 56(%rdx), %r14 movq 16(%rsi), %r8 movq (%rsi), %r15 movq 8(%rsi), %r11 xorl %eax, %eax subq (%rdx), %r15 sbbq 8(%rdx), %r11 sbbq 16(%rdx), %r8 movq 24(%rsi), %r12 sbbq 24(%rdx), %r12 movq %r15, (%rdi) movq 32(%rsi), %rbx sbbq 32(%rdx), %rbx movq %r11, 8(%rdi) movq 48(%rdx), %r15 movq 40(%rdx), %rdx movq %r8, 16(%rdi) movq 72(%rsi), %r8 movq %r12, 24(%rdi) movq 64(%rsi), %r11 movq %rbx, 32(%rdi) movq 40(%rsi), %rbp sbbq %rdx, %rbp movq 56(%rsi), %r12 movq 48(%rsi), %r13 sbbq %r15, %r13 sbbq %r14, %r12 sbbq %r10, %r11 sbbq %r9, %r8 movl $0, %edx sbbq $0, %rdx andl $1, %edx movq (%rcx), %rsi cmoveq %rax, %rsi testb %dl, %dl movq 16(%rcx), %rdx cmoveq %rax, %rdx movq 8(%rcx), %rbx cmoveq %rax, %rbx movq 32(%rcx), %r9 cmoveq %rax, %r9 cmovneq 24(%rcx), %rax addq %rbp, %rsi movq %rsi, 40(%rdi) adcq %r13, %rbx movq %rbx, 48(%rdi) adcq %r12, %rdx movq %rdx, 56(%rdi) adcq %r11, %rax movq %rax, 64(%rdi) adcq %r8, %r9 movq %r9, 72(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end79: .size mcl_fpDbl_sub5L, .Lfunc_end79-mcl_fpDbl_sub5L .globl mcl_fp_mulUnitPre6L .align 16, 0x90 .type mcl_fp_mulUnitPre6L,@function mcl_fp_mulUnitPre6L: # @mcl_fp_mulUnitPre6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rcx, %rax mulq 40(%rsi) movq %rdx, %r9 movq %rax, %r8 movq %rcx, %rax mulq 32(%rsi) movq %rdx, %r10 movq %rax, %r11 movq %rcx, %rax mulq 24(%rsi) movq %rdx, %r15 movq %rax, %r14 movq %rcx, %rax mulq 16(%rsi) movq %rdx, %r13 movq %rax, %r12 movq %rcx, %rax mulq 8(%rsi) movq %rdx, %rbx movq %rax, %rbp movq %rcx, %rax mulq (%rsi) movq %rax, (%rdi) addq %rbp, %rdx movq %rdx, 8(%rdi) adcq %r12, %rbx movq %rbx, 16(%rdi) adcq %r14, %r13 movq %r13, 24(%rdi) adcq %r11, %r15 movq %r15, 32(%rdi) adcq %r8, %r10 movq %r10, 40(%rdi) adcq $0, %r9 movq %r9, 48(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end80: .size mcl_fp_mulUnitPre6L, .Lfunc_end80-mcl_fp_mulUnitPre6L .globl mcl_fpDbl_mulPre6L .align 16, 0x90 .type mcl_fpDbl_mulPre6L,@function mcl_fpDbl_mulPre6L: # @mcl_fpDbl_mulPre6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, -16(%rsp) # 8-byte Spill movq %rsi, %r8 movq %rdi, -8(%rsp) # 8-byte Spill movq (%r8), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 8(%r8), %r13 movq %r13, -72(%rsp) # 8-byte Spill movq (%rdx), %rbx mulq %rbx movq %rdx, -32(%rsp) # 8-byte Spill movq 16(%r8), %rbp movq %rbp, -64(%rsp) # 8-byte Spill movq 24(%r8), %rsi movq %rsi, -48(%rsp) # 8-byte Spill movq 32(%r8), %r10 movq 40(%r8), %r11 movq %rax, (%rdi) movq %r11, %rax mulq %rbx movq %rdx, %rcx movq %rax, -40(%rsp) # 8-byte Spill movq %r10, %rax mulq %rbx movq %rdx, %r12 movq %rax, %rdi movq %rsi, %rax mulq %rbx movq %rdx, %r9 movq %rax, %r14 movq %rbp, %rax mulq %rbx movq %rdx, %rbp movq %rax, %r15 movq %r13, %rax mulq %rbx movq %rdx, %r13 movq %rax, %rsi addq -32(%rsp), %rsi # 8-byte Folded Reload adcq %r15, %r13 adcq %r14, %rbp adcq %rdi, %r9 adcq -40(%rsp), %r12 # 8-byte Folded Reload movq %r12, %rdi adcq $0, %rcx movq %rcx, -56(%rsp) # 8-byte Spill movq -16(%rsp), %r15 # 8-byte Reload movq 8(%r15), %rcx movq %r11, %rax mulq %rcx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r11 movq %r10, %rax mulq %rcx movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %r12 movq -48(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %r14 movq -64(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %rbx movq -72(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %r10 movq -24(%rsp), %rax # 8-byte Reload mulq %rcx addq %rsi, %rax movq -8(%rsp), %rcx # 8-byte Reload movq %rax, 8(%rcx) adcq %r13, %r10 adcq %rbp, %rbx adcq %r9, %r14 adcq %rdi, %r12 adcq -56(%rsp), %r11 # 8-byte Folded Reload sbbq %rdi, %rdi andl $1, %edi addq %rdx, %r10 adcq -72(%rsp), %rbx # 8-byte Folded Reload adcq -64(%rsp), %r14 # 8-byte Folded Reload adcq -48(%rsp), %r12 # 8-byte Folded Reload adcq -40(%rsp), %r11 # 8-byte Folded Reload movq %r11, -96(%rsp) # 8-byte Spill adcq -32(%rsp), %rdi # 8-byte Folded Reload movq 40(%r8), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 16(%r15), %rcx mulq %rcx movq %rax, -40(%rsp) # 8-byte Spill movq %rdx, -32(%rsp) # 8-byte Spill movq 32(%r8), %rax movq %rax, -48(%rsp) # 8-byte Spill mulq %rcx movq %rax, %r15 movq %rdx, -88(%rsp) # 8-byte Spill movq 24(%r8), %rax movq %rax, -56(%rsp) # 8-byte Spill mulq %rcx movq %rax, %r11 movq %rdx, -104(%rsp) # 8-byte Spill movq 16(%r8), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rcx movq %rax, %rbp movq %rdx, -112(%rsp) # 8-byte Spill movq (%r8), %rsi movq %rsi, -72(%rsp) # 8-byte Spill movq 8(%r8), %rax movq %rax, -80(%rsp) # 8-byte Spill mulq %rcx movq %rdx, %r13 movq %rax, %r9 movq %rsi, %rax mulq %rcx addq %r10, %rax movq -8(%rsp), %r10 # 8-byte Reload movq %rax, 16(%r10) adcq %rbx, %r9 adcq %r14, %rbp adcq %r12, %r11 adcq -96(%rsp), %r15 # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload adcq %rdi, %rax sbbq %rcx, %rcx andl $1, %ecx addq %rdx, %r9 adcq %r13, %rbp adcq -112(%rsp), %r11 # 8-byte Folded Reload adcq -104(%rsp), %r15 # 8-byte Folded Reload adcq -88(%rsp), %rax # 8-byte Folded Reload movq %rax, -40(%rsp) # 8-byte Spill adcq -32(%rsp), %rcx # 8-byte Folded Reload movq -16(%rsp), %rdi # 8-byte Reload movq 24(%rdi), %rbx movq -24(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -32(%rsp) # 8-byte Spill movq -48(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -24(%rsp) # 8-byte Spill movq -56(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %r14 movq -64(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %r12 movq -80(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, %rsi movq %rax, %r13 movq -72(%rsp), %rax # 8-byte Reload mulq %rbx addq %r9, %rax movq %rax, 24(%r10) adcq %rbp, %r13 adcq %r11, %r12 adcq %r15, %r14 movq -24(%rsp), %rbp # 8-byte Reload adcq -40(%rsp), %rbp # 8-byte Folded Reload movq -32(%rsp), %rax # 8-byte Reload adcq %rcx, %rax sbbq %r10, %r10 andl $1, %r10d addq %rdx, %r13 adcq %rsi, %r12 adcq -64(%rsp), %r14 # 8-byte Folded Reload adcq -56(%rsp), %rbp # 8-byte Folded Reload movq %rbp, -24(%rsp) # 8-byte Spill adcq -48(%rsp), %rax # 8-byte Folded Reload movq %rax, -32(%rsp) # 8-byte Spill adcq -88(%rsp), %r10 # 8-byte Folded Reload movq 40(%r8), %rax movq %rax, -72(%rsp) # 8-byte Spill movq 32(%rdi), %rcx movq 32(%r8), %rbx movq %rbx, -112(%rsp) # 8-byte Spill movq 24(%r8), %rsi movq %rsi, -64(%rsp) # 8-byte Spill movq 16(%r8), %rdi movq %rdi, -104(%rsp) # 8-byte Spill movq (%r8), %r15 movq 8(%r8), %r9 mulq %rcx movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %r11 movq %rbx, %rax mulq %rcx movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %r8 movq %rsi, %rax mulq %rcx movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %rbx movq %rdi, %rax mulq %rcx movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, %rdi movq %r9, %rax mulq %rcx movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, %rbp movq %r15, %rax mulq %rcx movq %rdx, -96(%rsp) # 8-byte Spill addq %r13, %rax movq -8(%rsp), %r13 # 8-byte Reload movq %rax, 32(%r13) adcq %r12, %rbp adcq %r14, %rdi adcq -24(%rsp), %rbx # 8-byte Folded Reload adcq -32(%rsp), %r8 # 8-byte Folded Reload adcq %r10, %r11 movq -16(%rsp), %rax # 8-byte Reload movq 40(%rax), %rcx sbbq %rsi, %rsi movq -72(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq -112(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r14 movq %r9, %rax mulq %rcx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r10 movq %r15, %rax mulq %rcx movq %rdx, %r12 movq %rax, %r9 movq -64(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %r15 movq -104(%rsp), %rax # 8-byte Reload mulq %rcx andl $1, %esi addq -96(%rsp), %rbp # 8-byte Folded Reload adcq -88(%rsp), %rdi # 8-byte Folded Reload adcq -80(%rsp), %rbx # 8-byte Folded Reload adcq -56(%rsp), %r8 # 8-byte Folded Reload adcq -48(%rsp), %r11 # 8-byte Folded Reload adcq -40(%rsp), %rsi # 8-byte Folded Reload addq %r9, %rbp movq %rbp, 40(%r13) adcq %r10, %rdi adcq %rax, %rbx adcq %r15, %r8 adcq %r14, %r11 adcq -72(%rsp), %rsi # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax addq %r12, %rdi movq %rdi, 48(%r13) adcq -32(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 56(%r13) adcq %rdx, %r8 movq %r8, 64(%r13) adcq -64(%rsp), %r11 # 8-byte Folded Reload movq %r11, 72(%r13) adcq -24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 80(%r13) adcq -16(%rsp), %rax # 8-byte Folded Reload movq %rax, 88(%r13) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end81: .size mcl_fpDbl_mulPre6L, .Lfunc_end81-mcl_fpDbl_mulPre6L .globl mcl_fpDbl_sqrPre6L .align 16, 0x90 .type mcl_fpDbl_sqrPre6L,@function mcl_fpDbl_sqrPre6L: # @mcl_fpDbl_sqrPre6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdi, -8(%rsp) # 8-byte Spill movq 16(%rsi), %r8 movq %r8, -56(%rsp) # 8-byte Spill movq 24(%rsi), %r10 movq %r10, -40(%rsp) # 8-byte Spill movq 32(%rsi), %r9 movq %r9, -32(%rsp) # 8-byte Spill movq 40(%rsi), %r11 movq (%rsi), %rcx movq 8(%rsi), %rbx movq %rcx, %rax mulq %rcx movq %rdx, %rbp movq %rax, (%rdi) movq %r11, %rax mulq %rcx movq %rdx, %rdi movq %rax, -16(%rsp) # 8-byte Spill movq %r9, %rax mulq %rcx movq %rdx, %r14 movq %rax, %r9 movq %r10, %rax mulq %rcx movq %rdx, %r12 movq %rax, %r10 movq %r8, %rax mulq %rcx movq %rdx, %r13 movq %rax, %r15 movq %rbx, %rax mulq %rcx movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r8 addq %r8, %rbp adcq %rdx, %r15 adcq %r10, %r13 adcq %r9, %r12 adcq -16(%rsp), %r14 # 8-byte Folded Reload adcq $0, %rdi movq %rdi, -48(%rsp) # 8-byte Spill movq %r11, %rax mulq %rbx movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %rcx movq -32(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r9 movq -40(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %r10 movq -56(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, %rdi movq %rax, %r11 movq %rbx, %rax mulq %rbx movq %rax, %rbx addq %r8, %rbp movq -8(%rsp), %rax # 8-byte Reload movq %rbp, 8(%rax) adcq %r15, %rbx adcq %r13, %r11 adcq %r12, %r10 adcq %r14, %r9 movq %rcx, %rax adcq -48(%rsp), %rax # 8-byte Folded Reload sbbq %rcx, %rcx andl $1, %ecx addq -24(%rsp), %rbx # 8-byte Folded Reload adcq %rdx, %r11 adcq %rdi, %r10 adcq -40(%rsp), %r9 # 8-byte Folded Reload adcq -32(%rsp), %rax # 8-byte Folded Reload movq %rax, -72(%rsp) # 8-byte Spill adcq -16(%rsp), %rcx # 8-byte Folded Reload movq 40(%rsi), %rax movq %rax, -16(%rsp) # 8-byte Spill movq 16(%rsi), %rdi mulq %rdi movq %rax, -80(%rsp) # 8-byte Spill movq %rdx, -40(%rsp) # 8-byte Spill movq 32(%rsi), %rax movq %rax, -32(%rsp) # 8-byte Spill mulq %rdi movq %rax, %r12 movq %rdx, -56(%rsp) # 8-byte Spill movq 24(%rsi), %rbp movq %rbp, %rax mulq %rdi movq %rax, %r14 movq %r14, -96(%rsp) # 8-byte Spill movq %rdx, -24(%rsp) # 8-byte Spill movq (%rsi), %r15 movq %r15, -48(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rdi movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, %r8 movq %r15, %rax mulq %rdi movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %r15 movq %rdi, %rax mulq %rdi movq %rax, %r13 addq %rbx, %r15 movq -8(%rsp), %rbx # 8-byte Reload movq %r15, 16(%rbx) adcq %r11, %r8 adcq %r10, %r13 adcq %r14, %r9 adcq -72(%rsp), %r12 # 8-byte Folded Reload movq -80(%rsp), %r14 # 8-byte Reload adcq %rcx, %r14 sbbq %rcx, %rcx andl $1, %ecx addq -104(%rsp), %r8 # 8-byte Folded Reload adcq -88(%rsp), %r13 # 8-byte Folded Reload adcq %rdx, %r9 adcq -24(%rsp), %r12 # 8-byte Folded Reload adcq -56(%rsp), %r14 # 8-byte Folded Reload adcq -40(%rsp), %rcx # 8-byte Folded Reload movq -16(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -16(%rsp) # 8-byte Spill movq -32(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %r10 movq -64(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, %r11 movq -48(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, %rdi movq %rbp, %rax mulq %rbp movq %rax, %r15 movq %rdx, -104(%rsp) # 8-byte Spill addq %r8, %rdi movq %rdi, 24(%rbx) adcq %r13, %r11 adcq -96(%rsp), %r9 # 8-byte Folded Reload adcq %r12, %r15 adcq %r14, %r10 movq -16(%rsp), %r12 # 8-byte Reload adcq %rcx, %r12 sbbq %rcx, %rcx movq (%rsi), %r8 andl $1, %ecx movq 8(%rsi), %rbx movq 40(%rsi), %rdi movq %rbx, %rax mulq %rdi movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, -48(%rsp) # 8-byte Spill movq %r8, %rax mulq %rdi movq %rax, -64(%rsp) # 8-byte Spill movq %rdx, -40(%rsp) # 8-byte Spill movq 32(%rsi), %rbp movq %rbx, %rax mulq %rbp movq %rdx, -96(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %r8, %rax mulq %rbp movq %rax, %r14 movq %rdx, -112(%rsp) # 8-byte Spill addq -88(%rsp), %r11 # 8-byte Folded Reload adcq -80(%rsp), %r9 # 8-byte Folded Reload adcq -24(%rsp), %r15 # 8-byte Folded Reload adcq -104(%rsp), %r10 # 8-byte Folded Reload adcq -72(%rsp), %r12 # 8-byte Folded Reload movq %r12, -16(%rsp) # 8-byte Spill adcq -56(%rsp), %rcx # 8-byte Folded Reload movq 24(%rsi), %rbx movq 16(%rsi), %r8 movq %rbx, %rax mulq %rdi movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rbx, %rax mulq %rbp movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %rsi movq %r8, %rax mulq %rdi movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %r8, %rax mulq %rbp movq %rdx, %rbx movq %rax, %r13 movq %rdi, %rax mulq %rbp movq %rdx, %r12 movq %rax, %r8 movq %rdi, %rax mulq %rdi movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %rdi movq %rbp, %rax mulq %rbp addq %r14, %r11 movq -8(%rsp), %r14 # 8-byte Reload movq %r11, 32(%r14) adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq %r15, %r13 adcq %r10, %rsi adcq -16(%rsp), %rax # 8-byte Folded Reload adcq %r8, %rcx sbbq %rbp, %rbp andl $1, %ebp addq -112(%rsp), %r9 # 8-byte Folded Reload adcq -96(%rsp), %r13 # 8-byte Folded Reload adcq %rbx, %rsi adcq -104(%rsp), %rax # 8-byte Folded Reload adcq %rdx, %rcx adcq %r12, %rbp addq -64(%rsp), %r9 # 8-byte Folded Reload movq %r14, %rbx movq %r9, 40(%rbx) adcq -48(%rsp), %r13 # 8-byte Folded Reload adcq -88(%rsp), %rsi # 8-byte Folded Reload adcq -80(%rsp), %rax # 8-byte Folded Reload adcq %r8, %rcx adcq %rdi, %rbp sbbq %rdx, %rdx andl $1, %edx addq -40(%rsp), %r13 # 8-byte Folded Reload movq %r13, 48(%rbx) adcq -32(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 56(%rbx) adcq -72(%rsp), %rax # 8-byte Folded Reload movq %rax, 64(%rbx) adcq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 72(%rbx) adcq %r12, %rbp movq %rbp, 80(%rbx) adcq -56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 88(%rbx) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end82: .size mcl_fpDbl_sqrPre6L, .Lfunc_end82-mcl_fpDbl_sqrPre6L .globl mcl_fp_mont6L .align 16, 0x90 .type mcl_fp_mont6L,@function mcl_fp_mont6L: # @mcl_fp_mont6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $56, %rsp movq %rdx, -8(%rsp) # 8-byte Spill movq %rdi, -96(%rsp) # 8-byte Spill movq 40(%rsi), %rax movq %rax, -16(%rsp) # 8-byte Spill movq (%rdx), %rbx mulq %rbx movq %rax, %r8 movq %rdx, %r14 movq 32(%rsi), %rax movq %rax, -24(%rsp) # 8-byte Spill mulq %rbx movq %rax, %r9 movq %rdx, %r15 movq 24(%rsi), %rax movq %rax, -56(%rsp) # 8-byte Spill movq 16(%rsi), %rbp movq %rbp, -48(%rsp) # 8-byte Spill movq (%rsi), %r12 movq %r12, -32(%rsp) # 8-byte Spill movq 8(%rsi), %rsi movq %rsi, -40(%rsp) # 8-byte Spill mulq %rbx movq %rdx, %rdi movq %rax, %r10 movq %rbp, %rax mulq %rbx movq %rdx, %rbp movq %rax, %r11 movq %rsi, %rax mulq %rbx movq %rdx, %rsi movq %rax, %r13 movq %r12, %rax mulq %rbx movq %rax, -120(%rsp) # 8-byte Spill addq %r13, %rdx movq %rdx, -112(%rsp) # 8-byte Spill adcq %r11, %rsi movq %rsi, -104(%rsp) # 8-byte Spill adcq %r10, %rbp movq %rbp, -88(%rsp) # 8-byte Spill adcq %r9, %rdi movq %rdi, -80(%rsp) # 8-byte Spill adcq %r8, %r15 movq %r15, -72(%rsp) # 8-byte Spill adcq $0, %r14 movq %r14, -64(%rsp) # 8-byte Spill movq -8(%rcx), %rdx movq %rdx, (%rsp) # 8-byte Spill movq %rax, %rdi imulq %rdx, %rdi movq (%rcx), %r9 movq %r9, 8(%rsp) # 8-byte Spill movq 40(%rcx), %rdx movq %rdx, 48(%rsp) # 8-byte Spill movq 32(%rcx), %rbp movq %rbp, 32(%rsp) # 8-byte Spill movq 24(%rcx), %rbx movq %rbx, 40(%rsp) # 8-byte Spill movq 16(%rcx), %rsi movq %rsi, 16(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, 24(%rsp) # 8-byte Spill movq %rdi, %rax mulq %rdx movq %rdx, %r11 movq %rax, -128(%rsp) # 8-byte Spill movq %rdi, %rax mulq %rbp movq %rdx, %r13 movq %rax, %r10 movq %rdi, %rax mulq %rbx movq %rdx, %rbp movq %rax, %r14 movq %rdi, %rax mulq %rsi movq %rdx, %rbx movq %rax, %r12 movq %rdi, %rax mulq %rcx movq %rdx, %r8 movq %rax, %r15 movq %rdi, %rax mulq %r9 movq %rdx, %r9 addq %r15, %r9 adcq %r12, %r8 adcq %r14, %rbx adcq %r10, %rbp adcq -128(%rsp), %r13 # 8-byte Folded Reload adcq $0, %r11 addq -120(%rsp), %rax # 8-byte Folded Reload adcq -112(%rsp), %r9 # 8-byte Folded Reload adcq -104(%rsp), %r8 # 8-byte Folded Reload adcq -88(%rsp), %rbx # 8-byte Folded Reload adcq -80(%rsp), %rbp # 8-byte Folded Reload adcq -72(%rsp), %r13 # 8-byte Folded Reload adcq -64(%rsp), %r11 # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax movq %rax, -80(%rsp) # 8-byte Spill movq -8(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdi movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %rcx movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r10 movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r12 movq %rdx, %rdi addq %r10, %rdi adcq %rcx, %rsi adcq -112(%rsp), %r15 # 8-byte Folded Reload adcq -104(%rsp), %r14 # 8-byte Folded Reload movq -72(%rsp), %rcx # 8-byte Reload adcq -88(%rsp), %rcx # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r9, %r12 adcq %r8, %rdi adcq %rbx, %rsi adcq %rbp, %r15 adcq %r13, %r14 adcq %r11, %rcx movq %rcx, -72(%rsp) # 8-byte Spill adcq -80(%rsp), %rax # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -88(%rsp) # 8-byte Spill movq %r12, %rbx imulq (%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbx, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -112(%rsp) # 8-byte Spill movq %rbx, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -120(%rsp) # 8-byte Spill movq %rbx, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r13 movq %rbx, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r11 movq %rbx, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 addq %r11, %r9 adcq %r13, %rbp adcq -120(%rsp), %rcx # 8-byte Folded Reload adcq -112(%rsp), %r10 # 8-byte Folded Reload adcq -104(%rsp), %r8 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r12, %rax adcq %rdi, %r9 adcq %rsi, %rbp adcq %r15, %rcx adcq %r14, %r10 adcq -72(%rsp), %r8 # 8-byte Folded Reload adcq -64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill adcq $0, -88(%rsp) # 8-byte Folded Spill movq -8(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdi movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %rbx movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %r11 movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r15 movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r13 movq %rdx, %rdi addq %r15, %rdi adcq %r11, %rsi adcq %rbx, %r12 adcq -112(%rsp), %r14 # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq -104(%rsp), %rdx # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r9, %r13 adcq %rbp, %rdi adcq %rcx, %rsi adcq %r10, %r12 adcq %r8, %r14 adcq -80(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill adcq -88(%rsp), %rax # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -88(%rsp) # 8-byte Spill movq %r13, %rbp imulq (%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r10 movq %rbp, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r8 movq %rbp, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 addq %r8, %r9 adcq %r10, %rcx adcq -120(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload adcq -104(%rsp), %r11 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r13, %rax adcq %rdi, %r9 adcq %rsi, %rcx adcq %r12, %rbx adcq %r14, %r15 adcq -72(%rsp), %r11 # 8-byte Folded Reload adcq -64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq -88(%rsp), %rbp # 8-byte Reload adcq $0, %rbp movq -8(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdi movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r8 movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r13 movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r12 movq %rdx, %rdi addq %r13, %rdi adcq %r8, %rsi adcq -112(%rsp), %r10 # 8-byte Folded Reload adcq -104(%rsp), %r14 # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq -88(%rsp), %rdx # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r9, %r12 adcq %rcx, %rdi adcq %rbx, %rsi adcq %r15, %r10 adcq %r11, %r14 adcq -80(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill adcq %rbp, %rax movq %rax, -64(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -88(%rsp) # 8-byte Spill movq %r12, %rbp imulq (%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r13 movq %rbp, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r9 movq %rbp, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r8 addq %r9, %r8 adcq %r13, %rcx adcq -120(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload adcq -104(%rsp), %r11 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r12, %rax adcq %rdi, %r8 adcq %rsi, %rcx adcq %r10, %rbx adcq %r14, %r15 adcq -72(%rsp), %r11 # 8-byte Folded Reload adcq -64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill adcq $0, -88(%rsp) # 8-byte Folded Spill movq -8(%rsp), %rax # 8-byte Reload movq 32(%rax), %rsi movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r12 movq %rsi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %r10 movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r9 movq %rdx, %r13 addq %r10, %r13 adcq %r12, %r14 adcq -120(%rsp), %rdi # 8-byte Folded Reload adcq -112(%rsp), %rbp # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq -104(%rsp), %rdx # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r8, %r9 adcq %rcx, %r13 adcq %rbx, %r14 adcq %r15, %rdi adcq %r11, %rbp adcq -80(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill adcq -88(%rsp), %rax # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -88(%rsp) # 8-byte Spill movq %r9, %rsi imulq (%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %rbx movq %rsi, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %rcx movq %rsi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r8 addq %rcx, %r8 adcq %rbx, %r12 adcq -120(%rsp), %r15 # 8-byte Folded Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload adcq -104(%rsp), %r10 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r9, %rax adcq %r13, %r8 adcq %r14, %r12 adcq %rdi, %r15 adcq %rbp, %r11 adcq -72(%rsp), %r10 # 8-byte Folded Reload movq %r10, -72(%rsp) # 8-byte Spill adcq -64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq -88(%rsp), %rbx # 8-byte Reload adcq $0, %rbx movq -8(%rsp), %rax # 8-byte Reload movq 40(%rax), %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, -24(%rsp) # 8-byte Spill movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, %r9 movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rsi movq %rcx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %rdi movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rax, %r13 movq %rdx, %rbp addq %rdi, %rbp adcq %rsi, %r14 adcq %r9, %r10 movq -8(%rsp), %rax # 8-byte Reload adcq -24(%rsp), %rax # 8-byte Folded Reload movq -16(%rsp), %rdx # 8-byte Reload adcq -88(%rsp), %rdx # 8-byte Folded Reload movq -64(%rsp), %rsi # 8-byte Reload adcq $0, %rsi addq %r8, %r13 movq %r13, -40(%rsp) # 8-byte Spill adcq %r12, %rbp adcq %r15, %r14 movq %r14, -24(%rsp) # 8-byte Spill adcq %r11, %r10 movq %r10, -32(%rsp) # 8-byte Spill adcq -72(%rsp), %rax # 8-byte Folded Reload movq %rax, -8(%rsp) # 8-byte Spill adcq -80(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -16(%rsp) # 8-byte Spill adcq %rbx, %rsi movq %rsi, -64(%rsp) # 8-byte Spill sbbq %rcx, %rcx movq (%rsp), %r9 # 8-byte Reload imulq %r13, %r9 andl $1, %ecx movq %r9, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, (%rsp) # 8-byte Spill movq %r9, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, -48(%rsp) # 8-byte Spill movq %r9, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, -56(%rsp) # 8-byte Spill movq %r9, %rax movq 8(%rsp), %r13 # 8-byte Reload mulq %r13 movq %rdx, %r15 movq %rax, %r12 movq %r9, %rax movq 16(%rsp), %r14 # 8-byte Reload mulq %r14 movq %rdx, %rsi movq %rax, %r11 movq %r9, %rax movq 24(%rsp), %r10 # 8-byte Reload mulq %r10 addq %r15, %rax adcq %r11, %rdx adcq -56(%rsp), %rsi # 8-byte Folded Reload adcq -48(%rsp), %rdi # 8-byte Folded Reload adcq (%rsp), %rbx # 8-byte Folded Reload adcq $0, %r8 addq -40(%rsp), %r12 # 8-byte Folded Reload adcq %rbp, %rax adcq -24(%rsp), %rdx # 8-byte Folded Reload adcq -32(%rsp), %rsi # 8-byte Folded Reload adcq -8(%rsp), %rdi # 8-byte Folded Reload adcq -16(%rsp), %rbx # 8-byte Folded Reload adcq -64(%rsp), %r8 # 8-byte Folded Reload adcq $0, %rcx movq %rax, %rbp subq %r13, %rbp movq %rdx, %r9 sbbq %r10, %r9 movq %rsi, %r10 sbbq %r14, %r10 movq %rdi, %r11 sbbq 40(%rsp), %r11 # 8-byte Folded Reload movq %rbx, %r14 sbbq 32(%rsp), %r14 # 8-byte Folded Reload movq %r8, %r15 sbbq 48(%rsp), %r15 # 8-byte Folded Reload sbbq $0, %rcx andl $1, %ecx cmovneq %rdi, %r11 testb %cl, %cl cmovneq %rax, %rbp movq -96(%rsp), %rax # 8-byte Reload movq %rbp, (%rax) cmovneq %rdx, %r9 movq %r9, 8(%rax) cmovneq %rsi, %r10 movq %r10, 16(%rax) movq %r11, 24(%rax) cmovneq %rbx, %r14 movq %r14, 32(%rax) cmovneq %r8, %r15 movq %r15, 40(%rax) addq $56, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end83: .size mcl_fp_mont6L, .Lfunc_end83-mcl_fp_mont6L .globl mcl_fp_montNF6L .align 16, 0x90 .type mcl_fp_montNF6L,@function mcl_fp_montNF6L: # @mcl_fp_montNF6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $40, %rsp movq %rdx, -8(%rsp) # 8-byte Spill movq %rdi, -88(%rsp) # 8-byte Spill movq 40(%rsi), %rax movq %rax, -16(%rsp) # 8-byte Spill movq (%rdx), %rbx mulq %rbx movq %rax, 32(%rsp) # 8-byte Spill movq %rdx, %r13 movq 32(%rsi), %rax movq %rax, -24(%rsp) # 8-byte Spill mulq %rbx movq %rax, %r10 movq %rdx, %r9 movq 24(%rsi), %rax movq %rax, -72(%rsp) # 8-byte Spill movq 16(%rsi), %rbp movq %rbp, -64(%rsp) # 8-byte Spill movq (%rsi), %rdi movq %rdi, -48(%rsp) # 8-byte Spill movq 8(%rsi), %rsi movq %rsi, -56(%rsp) # 8-byte Spill mulq %rbx movq %rdx, %r11 movq %rax, %r8 movq %rbp, %rax mulq %rbx movq %rdx, %r14 movq %rax, %r15 movq %rsi, %rax mulq %rbx movq %rdx, %r12 movq %rax, %rbp movq %rdi, %rax mulq %rbx movq %rax, -104(%rsp) # 8-byte Spill movq %rdx, %rbx addq %rbp, %rbx adcq %r15, %r12 adcq %r8, %r14 adcq %r10, %r11 adcq 32(%rsp), %r9 # 8-byte Folded Reload movq %r9, -96(%rsp) # 8-byte Spill adcq $0, %r13 movq %r13, -80(%rsp) # 8-byte Spill movq -8(%rcx), %rdx movq %rdx, (%rsp) # 8-byte Spill movq %rax, %r9 imulq %rdx, %r9 movq (%rcx), %r8 movq %r8, 8(%rsp) # 8-byte Spill movq 40(%rcx), %rdx movq %rdx, 32(%rsp) # 8-byte Spill movq 32(%rcx), %rsi movq %rsi, 24(%rsp) # 8-byte Spill movq 24(%rcx), %rbp movq %rbp, 16(%rsp) # 8-byte Spill movq 16(%rcx), %rdi movq %rdi, -40(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -32(%rsp) # 8-byte Spill movq %r9, %rax mulq %rdx movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r10 movq %r9, %rax mulq %rsi movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r15 movq %r9, %rax mulq %rbp movq %rdx, -128(%rsp) # 8-byte Spill movq %rax, %rsi movq %r9, %rax mulq %rdi movq %rdx, %r13 movq %rax, %rdi movq %r9, %rax mulq %rcx movq %rdx, %rcx movq %rax, %rbp movq %r9, %rax mulq %r8 addq -104(%rsp), %rax # 8-byte Folded Reload adcq %rbx, %rbp adcq %r12, %rdi adcq %r14, %rsi adcq %r11, %r15 adcq -96(%rsp), %r10 # 8-byte Folded Reload movq -80(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rdx, %rbp adcq %rcx, %rdi adcq %r13, %rsi adcq -128(%rsp), %r15 # 8-byte Folded Reload adcq -120(%rsp), %r10 # 8-byte Folded Reload adcq -112(%rsp), %rax # 8-byte Folded Reload movq %rax, -80(%rsp) # 8-byte Spill movq -8(%rsp), %rax # 8-byte Reload movq 8(%rax), %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -96(%rsp) # 8-byte Spill movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -104(%rsp) # 8-byte Spill movq %rcx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r8 movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rax, %r13 movq %rdx, %rcx addq %r8, %rcx adcq -120(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %r12 # 8-byte Folded Reload adcq -104(%rsp), %r11 # 8-byte Folded Reload adcq -96(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r14 addq %rbp, %r13 adcq %rdi, %rcx adcq %rsi, %rbx adcq %r15, %r12 adcq %r10, %r11 adcq -80(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r14 movq %r13, %rsi imulq (%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, -96(%rsp) # 8-byte Spill movq %rax, %rdi movq %rsi, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, %rbp movq %rsi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %r8 movq %rsi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r10 movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r15 movq %rsi, %rax mulq 8(%rsp) # 8-byte Folded Reload addq %r13, %rax adcq %rcx, %r15 adcq %rbx, %r10 adcq %r12, %r8 adcq %r11, %rbp adcq %r9, %rdi adcq $0, %r14 addq %rdx, %r15 adcq -120(%rsp), %r10 # 8-byte Folded Reload adcq -112(%rsp), %r8 # 8-byte Folded Reload adcq -104(%rsp), %rbp # 8-byte Folded Reload adcq -80(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -80(%rsp) # 8-byte Spill adcq -96(%rsp), %r14 # 8-byte Folded Reload movq -8(%rsp), %rax # 8-byte Reload movq 16(%rax), %rsi movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -96(%rsp) # 8-byte Spill movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %rdi movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rax, %r9 movq %rdx, %rsi addq %rdi, %rsi adcq -120(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %r12 # 8-byte Folded Reload adcq -104(%rsp), %rcx # 8-byte Folded Reload adcq -96(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r13 addq %r15, %r9 adcq %r10, %rsi adcq %r8, %rbx adcq %rbp, %r12 adcq -80(%rsp), %rcx # 8-byte Folded Reload adcq %r14, %r11 adcq $0, %r13 movq %r9, %r8 imulq (%rsp), %r8 # 8-byte Folded Reload movq %r8, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %r8, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r15 movq %r8, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -96(%rsp) # 8-byte Spill movq %rax, %r10 movq %r8, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r14 movq %r8, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %rdi movq %r8, %rax mulq 8(%rsp) # 8-byte Folded Reload addq %r9, %rax adcq %rsi, %rdi adcq %rbx, %r14 adcq %r12, %r10 adcq %rcx, %r15 movq -80(%rsp), %rax # 8-byte Reload adcq %r11, %rax adcq $0, %r13 addq %rdx, %rdi adcq %rbp, %r14 adcq -120(%rsp), %r10 # 8-byte Folded Reload adcq -96(%rsp), %r15 # 8-byte Folded Reload movq %r15, -96(%rsp) # 8-byte Spill adcq -112(%rsp), %rax # 8-byte Folded Reload movq %rax, -80(%rsp) # 8-byte Spill adcq -104(%rsp), %r13 # 8-byte Folded Reload movq -8(%rsp), %rax # 8-byte Reload movq 24(%rax), %rbp movq %rbp, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r12 movq %rbp, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r9 movq %rbp, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rax, %r8 movq %rdx, %rbp addq %r9, %rbp adcq %r12, %rbx adcq -120(%rsp), %rsi # 8-byte Folded Reload adcq -112(%rsp), %rcx # 8-byte Folded Reload adcq -104(%rsp), %r11 # 8-byte Folded Reload adcq $0, %r15 addq %rdi, %r8 adcq %r14, %rbp adcq %r10, %rbx adcq -96(%rsp), %rsi # 8-byte Folded Reload adcq -80(%rsp), %rcx # 8-byte Folded Reload adcq %r13, %r11 adcq $0, %r15 movq %r8, %r14 imulq (%rsp), %r14 # 8-byte Folded Reload movq %r14, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %r9 movq %r14, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, %r13 movq %r14, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -96(%rsp) # 8-byte Spill movq %rax, %r10 movq %r14, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r12 movq %r14, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %rdi movq %r14, %rax mulq 8(%rsp) # 8-byte Folded Reload addq %r8, %rax adcq %rbp, %rdi adcq %rbx, %r12 adcq %rsi, %r10 adcq %rcx, %r13 adcq %r11, %r9 adcq $0, %r15 addq %rdx, %rdi adcq -120(%rsp), %r12 # 8-byte Folded Reload adcq -112(%rsp), %r10 # 8-byte Folded Reload adcq -96(%rsp), %r13 # 8-byte Folded Reload movq %r13, -96(%rsp) # 8-byte Spill adcq -80(%rsp), %r9 # 8-byte Folded Reload movq %r9, -80(%rsp) # 8-byte Spill adcq -104(%rsp), %r15 # 8-byte Folded Reload movq -8(%rsp), %rax # 8-byte Reload movq 32(%rax), %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -104(%rsp) # 8-byte Spill movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, -128(%rsp) # 8-byte Spill movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r13 movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rax, %r11 movq %rdx, %rbp addq %r13, %rbp adcq -128(%rsp), %rbx # 8-byte Folded Reload adcq -120(%rsp), %rsi # 8-byte Folded Reload adcq -112(%rsp), %r8 # 8-byte Folded Reload adcq -104(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r14 addq %rdi, %r11 adcq %r12, %rbp adcq %r10, %rbx adcq -96(%rsp), %rsi # 8-byte Folded Reload adcq -80(%rsp), %r8 # 8-byte Folded Reload adcq %r15, %r9 adcq $0, %r14 movq %r11, %rcx imulq (%rsp), %rcx # 8-byte Folded Reload movq %rcx, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rcx, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r12 movq %rcx, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -96(%rsp) # 8-byte Spill movq %rax, %r10 movq %rcx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %r15 movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %rdi movq %rcx, %rax mulq 8(%rsp) # 8-byte Folded Reload addq %r11, %rax adcq %rbp, %rdi adcq %rbx, %r15 adcq %rsi, %r10 adcq %r8, %r12 movq -80(%rsp), %rcx # 8-byte Reload adcq %r9, %rcx adcq $0, %r14 addq %rdx, %rdi adcq %r13, %r15 adcq -104(%rsp), %r10 # 8-byte Folded Reload movq %r10, -104(%rsp) # 8-byte Spill adcq -96(%rsp), %r12 # 8-byte Folded Reload movq %r12, -96(%rsp) # 8-byte Spill adcq -120(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -80(%rsp) # 8-byte Spill adcq -112(%rsp), %r14 # 8-byte Folded Reload movq -8(%rsp), %rax # 8-byte Reload movq 40(%rax), %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, -8(%rsp) # 8-byte Spill movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -16(%rsp) # 8-byte Spill movq %rcx, %rax mulq -72(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -24(%rsp) # 8-byte Spill movq %rcx, %rax mulq -64(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %rbp movq %rcx, %rax mulq -56(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rsi movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rax, %r11 movq %rdx, %r8 addq %rsi, %r8 adcq %rbp, %r10 adcq -24(%rsp), %r13 # 8-byte Folded Reload adcq -16(%rsp), %r12 # 8-byte Folded Reload adcq -8(%rsp), %r9 # 8-byte Folded Reload adcq $0, %rbx addq %rdi, %r11 adcq %r15, %r8 adcq -104(%rsp), %r10 # 8-byte Folded Reload adcq -96(%rsp), %r13 # 8-byte Folded Reload adcq -80(%rsp), %r12 # 8-byte Folded Reload adcq %r14, %r9 movq %r9, -16(%rsp) # 8-byte Spill adcq $0, %rbx movq (%rsp), %r9 # 8-byte Reload imulq %r11, %r9 movq %r9, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, (%rsp) # 8-byte Spill movq %rax, %rsi movq %r9, %rax mulq 24(%rsp) # 8-byte Folded Reload movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, %rdi movq %r9, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %rbp movq %r9, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %r14 movq %r9, %rax movq -40(%rsp), %r15 # 8-byte Reload mulq %r15 movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %rcx movq %r9, %rax movq -32(%rsp), %r9 # 8-byte Reload mulq %r9 addq %r11, %r14 adcq %r8, %rax adcq %r10, %rcx adcq %r13, %rbp adcq %r12, %rdi adcq -16(%rsp), %rsi # 8-byte Folded Reload adcq $0, %rbx addq -48(%rsp), %rax # 8-byte Folded Reload adcq %rdx, %rcx adcq -56(%rsp), %rbp # 8-byte Folded Reload adcq -24(%rsp), %rdi # 8-byte Folded Reload adcq -8(%rsp), %rsi # 8-byte Folded Reload adcq (%rsp), %rbx # 8-byte Folded Reload movq %rax, %r14 subq 8(%rsp), %r14 # 8-byte Folded Reload movq %rcx, %r8 sbbq %r9, %r8 movq %rbp, %r9 sbbq %r15, %r9 movq %rdi, %r10 sbbq 16(%rsp), %r10 # 8-byte Folded Reload movq %rsi, %r11 sbbq 24(%rsp), %r11 # 8-byte Folded Reload movq %rbx, %r15 sbbq 32(%rsp), %r15 # 8-byte Folded Reload movq %r15, %rdx sarq $63, %rdx cmovsq %rax, %r14 movq -88(%rsp), %rax # 8-byte Reload movq %r14, (%rax) cmovsq %rcx, %r8 movq %r8, 8(%rax) cmovsq %rbp, %r9 movq %r9, 16(%rax) cmovsq %rdi, %r10 movq %r10, 24(%rax) cmovsq %rsi, %r11 movq %r11, 32(%rax) cmovsq %rbx, %r15 movq %r15, 40(%rax) addq $40, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end84: .size mcl_fp_montNF6L, .Lfunc_end84-mcl_fp_montNF6L .globl mcl_fp_montRed6L .align 16, 0x90 .type mcl_fp_montRed6L,@function mcl_fp_montRed6L: # @mcl_fp_montRed6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $16, %rsp movq %rdx, %rcx movq %rdi, -104(%rsp) # 8-byte Spill movq -8(%rcx), %rax movq %rax, -32(%rsp) # 8-byte Spill movq (%rcx), %r11 movq %r11, -24(%rsp) # 8-byte Spill movq (%rsi), %r9 movq %r9, %rbp imulq %rax, %rbp movq 40(%rcx), %rdx movq %rdx, -40(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rax, %r12 movq %rdx, -72(%rsp) # 8-byte Spill movq 32(%rcx), %rdx movq %rdx, 8(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rax, %r15 movq %rdx, %r8 movq 24(%rcx), %rdx movq %rdx, (%rsp) # 8-byte Spill movq 16(%rcx), %rdi movq %rdi, -8(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, -16(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rdx, %r10 movq %rax, %r14 movq %rbp, %rax mulq %rdi movq %rdx, %r13 movq %rax, %rbx movq %rbp, %rax mulq %rcx movq %rdx, %rcx movq %rax, %rdi movq %rbp, %rax mulq %r11 movq %rdx, %rbp addq %rdi, %rbp adcq %rbx, %rcx adcq %r14, %r13 adcq %r15, %r10 adcq %r12, %r8 movq -72(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r9, %rax adcq 8(%rsi), %rbp adcq 16(%rsi), %rcx adcq 24(%rsi), %r13 adcq 32(%rsi), %r10 adcq 40(%rsi), %r8 movq %r8, -112(%rsp) # 8-byte Spill adcq 48(%rsi), %rdx movq %rdx, -72(%rsp) # 8-byte Spill movq 88(%rsi), %rax movq 80(%rsi), %rdx movq 72(%rsi), %rdi movq 64(%rsi), %rbx movq 56(%rsi), %r15 adcq $0, %r15 adcq $0, %rbx movq %rbx, -96(%rsp) # 8-byte Spill adcq $0, %rdi movq %rdi, -64(%rsp) # 8-byte Spill adcq $0, %rdx movq %rdx, -56(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, -48(%rsp) # 8-byte Spill sbbq %r12, %r12 andl $1, %r12d movq %rbp, %rdi imulq -32(%rsp), %rdi # 8-byte Folded Reload movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -128(%rsp) # 8-byte Spill movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %r8 movq %rdi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r9 movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r11 movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rdi addq %r11, %rdi adcq %r9, %rsi adcq %r8, %rbx adcq -128(%rsp), %r14 # 8-byte Folded Reload movq -88(%rsp), %r8 # 8-byte Reload adcq -120(%rsp), %r8 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %rbp, %rax adcq %rcx, %rdi adcq %r13, %rsi adcq %r10, %rbx adcq -112(%rsp), %r14 # 8-byte Folded Reload adcq -72(%rsp), %r8 # 8-byte Folded Reload movq %r8, -88(%rsp) # 8-byte Spill adcq %r15, %rdx movq %rdx, -80(%rsp) # 8-byte Spill adcq $0, -96(%rsp) # 8-byte Folded Spill adcq $0, -64(%rsp) # 8-byte Folded Spill adcq $0, -56(%rsp) # 8-byte Folded Spill adcq $0, -48(%rsp) # 8-byte Folded Spill adcq $0, %r12 movq %rdi, %rcx imulq -32(%rsp), %rcx # 8-byte Folded Reload movq %rcx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -128(%rsp) # 8-byte Spill movq %rcx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r8 movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r10 movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r9 addq %r10, %r9 adcq %r8, %rbp adcq -128(%rsp), %r13 # 8-byte Folded Reload adcq -120(%rsp), %r11 # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload movq -72(%rsp), %rcx # 8-byte Reload adcq $0, %rcx addq %rdi, %rax adcq %rsi, %r9 adcq %rbx, %rbp adcq %r14, %r13 adcq -88(%rsp), %r11 # 8-byte Folded Reload adcq -80(%rsp), %r15 # 8-byte Folded Reload adcq -96(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -72(%rsp) # 8-byte Spill adcq $0, -64(%rsp) # 8-byte Folded Spill adcq $0, -56(%rsp) # 8-byte Folded Spill adcq $0, -48(%rsp) # 8-byte Folded Spill adcq $0, %r12 movq %r9, %rsi imulq -32(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -96(%rsp) # 8-byte Spill movq %rsi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r10 movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %rbx movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rdi addq %rbx, %rdi adcq %r10, %rcx adcq -120(%rsp), %r8 # 8-byte Folded Reload adcq -112(%rsp), %r14 # 8-byte Folded Reload movq -88(%rsp), %rsi # 8-byte Reload adcq -96(%rsp), %rsi # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r9, %rax adcq %rbp, %rdi adcq %r13, %rcx adcq %r11, %r8 adcq %r15, %r14 adcq -72(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -88(%rsp) # 8-byte Spill adcq -64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill adcq $0, -56(%rsp) # 8-byte Folded Spill movq -48(%rsp), %rbp # 8-byte Reload adcq $0, %rbp adcq $0, %r12 movq %rdi, %rsi imulq -32(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax movq -40(%rsp), %r11 # 8-byte Reload mulq %r11 movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -48(%rsp) # 8-byte Spill movq %rsi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -96(%rsp) # 8-byte Spill movq %rsi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r15 movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rbx movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rsi addq %rbx, %rsi adcq %r15, %r10 adcq -112(%rsp), %r13 # 8-byte Folded Reload adcq -96(%rsp), %r9 # 8-byte Folded Reload movq -72(%rsp), %rbx # 8-byte Reload adcq -48(%rsp), %rbx # 8-byte Folded Reload movq -64(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %rdi, %rax adcq %rcx, %rsi adcq %r8, %r10 adcq %r14, %r13 adcq -88(%rsp), %r9 # 8-byte Folded Reload adcq -80(%rsp), %rbx # 8-byte Folded Reload movq %rbx, -72(%rsp) # 8-byte Spill adcq -56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill adcq $0, %rbp movq %rbp, -48(%rsp) # 8-byte Spill adcq $0, %r12 movq -32(%rsp), %r8 # 8-byte Reload imulq %rsi, %r8 movq %r8, %rax mulq %r11 movq %rdx, %rdi movq %rax, -32(%rsp) # 8-byte Spill movq %r8, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -56(%rsp) # 8-byte Spill movq %r8, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, -80(%rsp) # 8-byte Spill movq %r8, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r11 movq %r8, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r14 movq %r8, %rax movq -24(%rsp), %r8 # 8-byte Reload mulq %r8 addq %r14, %rdx adcq %r11, %rbp adcq -80(%rsp), %rbx # 8-byte Folded Reload adcq -56(%rsp), %rcx # 8-byte Folded Reload adcq -32(%rsp), %r15 # 8-byte Folded Reload adcq $0, %rdi addq %rsi, %rax adcq %r10, %rdx adcq %r13, %rbp adcq %r9, %rbx adcq -72(%rsp), %rcx # 8-byte Folded Reload adcq -64(%rsp), %r15 # 8-byte Folded Reload adcq -48(%rsp), %rdi # 8-byte Folded Reload adcq $0, %r12 movq %rdx, %rax subq %r8, %rax movq %rbp, %rsi sbbq -16(%rsp), %rsi # 8-byte Folded Reload movq %rbx, %r9 sbbq -8(%rsp), %r9 # 8-byte Folded Reload movq %rcx, %r10 sbbq (%rsp), %r10 # 8-byte Folded Reload movq %r15, %r11 sbbq 8(%rsp), %r11 # 8-byte Folded Reload movq %rdi, %r14 sbbq -40(%rsp), %r14 # 8-byte Folded Reload sbbq $0, %r12 andl $1, %r12d cmovneq %rdi, %r14 testb %r12b, %r12b cmovneq %rdx, %rax movq -104(%rsp), %rdx # 8-byte Reload movq %rax, (%rdx) cmovneq %rbp, %rsi movq %rsi, 8(%rdx) cmovneq %rbx, %r9 movq %r9, 16(%rdx) cmovneq %rcx, %r10 movq %r10, 24(%rdx) cmovneq %r15, %r11 movq %r11, 32(%rdx) movq %r14, 40(%rdx) addq $16, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end85: .size mcl_fp_montRed6L, .Lfunc_end85-mcl_fp_montRed6L .globl mcl_fp_addPre6L .align 16, 0x90 .type mcl_fp_addPre6L,@function mcl_fp_addPre6L: # @mcl_fp_addPre6L # BB#0: pushq %r14 pushq %rbx movq 40(%rdx), %r8 movq 40(%rsi), %r11 movq 32(%rdx), %r9 movq 24(%rdx), %r10 movq 24(%rsi), %rax movq 32(%rsi), %r14 movq 16(%rdx), %rbx movq (%rdx), %rcx movq 8(%rdx), %rdx addq (%rsi), %rcx adcq 8(%rsi), %rdx adcq 16(%rsi), %rbx movq %rcx, (%rdi) movq %rdx, 8(%rdi) movq %rbx, 16(%rdi) adcq %r10, %rax movq %rax, 24(%rdi) adcq %r9, %r14 movq %r14, 32(%rdi) adcq %r8, %r11 movq %r11, 40(%rdi) sbbq %rax, %rax andl $1, %eax popq %rbx popq %r14 retq .Lfunc_end86: .size mcl_fp_addPre6L, .Lfunc_end86-mcl_fp_addPre6L .globl mcl_fp_subPre6L .align 16, 0x90 .type mcl_fp_subPre6L,@function mcl_fp_subPre6L: # @mcl_fp_subPre6L # BB#0: pushq %r15 pushq %r14 pushq %rbx movq 40(%rdx), %r8 movq 40(%rsi), %r9 movq 32(%rsi), %r10 movq 24(%rsi), %r11 movq 16(%rsi), %rcx movq (%rsi), %rbx movq 8(%rsi), %rsi xorl %eax, %eax subq (%rdx), %rbx sbbq 8(%rdx), %rsi movq 24(%rdx), %r14 movq 32(%rdx), %r15 sbbq 16(%rdx), %rcx movq %rbx, (%rdi) movq %rsi, 8(%rdi) movq %rcx, 16(%rdi) sbbq %r14, %r11 movq %r11, 24(%rdi) sbbq %r15, %r10 movq %r10, 32(%rdi) sbbq %r8, %r9 movq %r9, 40(%rdi) sbbq $0, %rax andl $1, %eax popq %rbx popq %r14 popq %r15 retq .Lfunc_end87: .size mcl_fp_subPre6L, .Lfunc_end87-mcl_fp_subPre6L .globl mcl_fp_shr1_6L .align 16, 0x90 .type mcl_fp_shr1_6L,@function mcl_fp_shr1_6L: # @mcl_fp_shr1_6L # BB#0: movq 40(%rsi), %r8 movq 32(%rsi), %r9 movq 24(%rsi), %rdx movq 16(%rsi), %rax movq (%rsi), %rcx movq 8(%rsi), %rsi shrdq $1, %rsi, %rcx movq %rcx, (%rdi) shrdq $1, %rax, %rsi movq %rsi, 8(%rdi) shrdq $1, %rdx, %rax movq %rax, 16(%rdi) shrdq $1, %r9, %rdx movq %rdx, 24(%rdi) shrdq $1, %r8, %r9 movq %r9, 32(%rdi) shrq %r8 movq %r8, 40(%rdi) retq .Lfunc_end88: .size mcl_fp_shr1_6L, .Lfunc_end88-mcl_fp_shr1_6L .globl mcl_fp_add6L .align 16, 0x90 .type mcl_fp_add6L,@function mcl_fp_add6L: # @mcl_fp_add6L # BB#0: pushq %r15 pushq %r14 pushq %rbx movq 40(%rdx), %r14 movq 40(%rsi), %r8 movq 32(%rdx), %r15 movq 24(%rdx), %rbx movq 24(%rsi), %r10 movq 32(%rsi), %r9 movq 16(%rdx), %r11 movq (%rdx), %rax movq 8(%rdx), %rdx addq (%rsi), %rax adcq 8(%rsi), %rdx adcq 16(%rsi), %r11 movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r11, 16(%rdi) adcq %rbx, %r10 movq %r10, 24(%rdi) adcq %r15, %r9 movq %r9, 32(%rdi) adcq %r14, %r8 movq %r8, 40(%rdi) sbbq %rsi, %rsi andl $1, %esi subq (%rcx), %rax sbbq 8(%rcx), %rdx sbbq 16(%rcx), %r11 sbbq 24(%rcx), %r10 sbbq 32(%rcx), %r9 sbbq 40(%rcx), %r8 sbbq $0, %rsi testb $1, %sil jne .LBB89_2 # BB#1: # %nocarry movq %rax, (%rdi) movq %rdx, 8(%rdi) movq %r11, 16(%rdi) movq %r10, 24(%rdi) movq %r9, 32(%rdi) movq %r8, 40(%rdi) .LBB89_2: # %carry popq %rbx popq %r14 popq %r15 retq .Lfunc_end89: .size mcl_fp_add6L, .Lfunc_end89-mcl_fp_add6L .globl mcl_fp_addNF6L .align 16, 0x90 .type mcl_fp_addNF6L,@function mcl_fp_addNF6L: # @mcl_fp_addNF6L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 40(%rdx), %r8 movq 32(%rdx), %r9 movq 24(%rdx), %r10 movq 16(%rdx), %r11 movq (%rdx), %r15 movq 8(%rdx), %r14 addq (%rsi), %r15 adcq 8(%rsi), %r14 adcq 16(%rsi), %r11 adcq 24(%rsi), %r10 adcq 32(%rsi), %r9 adcq 40(%rsi), %r8 movq %r15, %rsi subq (%rcx), %rsi movq %r14, %rbx sbbq 8(%rcx), %rbx movq %r11, %rdx sbbq 16(%rcx), %rdx movq %r10, %r13 sbbq 24(%rcx), %r13 movq %r9, %r12 sbbq 32(%rcx), %r12 movq %r8, %rax sbbq 40(%rcx), %rax movq %rax, %rcx sarq $63, %rcx cmovsq %r15, %rsi movq %rsi, (%rdi) cmovsq %r14, %rbx movq %rbx, 8(%rdi) cmovsq %r11, %rdx movq %rdx, 16(%rdi) cmovsq %r10, %r13 movq %r13, 24(%rdi) cmovsq %r9, %r12 movq %r12, 32(%rdi) cmovsq %r8, %rax movq %rax, 40(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end90: .size mcl_fp_addNF6L, .Lfunc_end90-mcl_fp_addNF6L .globl mcl_fp_sub6L .align 16, 0x90 .type mcl_fp_sub6L,@function mcl_fp_sub6L: # @mcl_fp_sub6L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq 40(%rdx), %r14 movq 40(%rsi), %r8 movq 32(%rsi), %r9 movq 24(%rsi), %r10 movq 16(%rsi), %r11 movq (%rsi), %rax movq 8(%rsi), %rsi xorl %ebx, %ebx subq (%rdx), %rax sbbq 8(%rdx), %rsi movq 24(%rdx), %r15 movq 32(%rdx), %r12 sbbq 16(%rdx), %r11 movq %rax, (%rdi) movq %rsi, 8(%rdi) movq %r11, 16(%rdi) sbbq %r15, %r10 movq %r10, 24(%rdi) sbbq %r12, %r9 movq %r9, 32(%rdi) sbbq %r14, %r8 movq %r8, 40(%rdi) sbbq $0, %rbx testb $1, %bl je .LBB91_2 # BB#1: # %carry movq 40(%rcx), %r14 movq 32(%rcx), %r15 movq 24(%rcx), %r12 movq 8(%rcx), %rbx movq 16(%rcx), %rdx addq (%rcx), %rax movq %rax, (%rdi) adcq %rsi, %rbx movq %rbx, 8(%rdi) adcq %r11, %rdx movq %rdx, 16(%rdi) adcq %r10, %r12 movq %r12, 24(%rdi) adcq %r9, %r15 movq %r15, 32(%rdi) adcq %r8, %r14 movq %r14, 40(%rdi) .LBB91_2: # %nocarry popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end91: .size mcl_fp_sub6L, .Lfunc_end91-mcl_fp_sub6L .globl mcl_fp_subNF6L .align 16, 0x90 .type mcl_fp_subNF6L,@function mcl_fp_subNF6L: # @mcl_fp_subNF6L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 40(%rsi), %r15 movq 32(%rsi), %r8 movq 24(%rsi), %r9 movq 16(%rsi), %r10 movq (%rsi), %r11 movq 8(%rsi), %r14 subq (%rdx), %r11 sbbq 8(%rdx), %r14 sbbq 16(%rdx), %r10 sbbq 24(%rdx), %r9 sbbq 32(%rdx), %r8 sbbq 40(%rdx), %r15 movq %r15, %rdx sarq $63, %rdx movq %rdx, %rbx addq %rbx, %rbx movq %rdx, %rsi adcq %rsi, %rsi andq 8(%rcx), %rsi movq %r15, %rax shrq $63, %rax orq %rbx, %rax andq (%rcx), %rax movq 40(%rcx), %r12 andq %rdx, %r12 movq 32(%rcx), %r13 andq %rdx, %r13 movq 24(%rcx), %rbx andq %rdx, %rbx andq 16(%rcx), %rdx addq %r11, %rax movq %rax, (%rdi) adcq %r14, %rsi movq %rsi, 8(%rdi) adcq %r10, %rdx movq %rdx, 16(%rdi) adcq %r9, %rbx movq %rbx, 24(%rdi) adcq %r8, %r13 movq %r13, 32(%rdi) adcq %r15, %r12 movq %r12, 40(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end92: .size mcl_fp_subNF6L, .Lfunc_end92-mcl_fp_subNF6L .globl mcl_fpDbl_add6L .align 16, 0x90 .type mcl_fpDbl_add6L,@function mcl_fpDbl_add6L: # @mcl_fpDbl_add6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 88(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 80(%rdx), %rax movq %rax, -16(%rsp) # 8-byte Spill movq 72(%rdx), %r14 movq 64(%rdx), %r15 movq 24(%rsi), %rbp movq 32(%rsi), %r13 movq 16(%rdx), %r12 movq (%rdx), %rbx movq 8(%rdx), %rax addq (%rsi), %rbx adcq 8(%rsi), %rax adcq 16(%rsi), %r12 adcq 24(%rdx), %rbp adcq 32(%rdx), %r13 movq 56(%rdx), %r11 movq 48(%rdx), %r9 movq 40(%rdx), %rdx movq %rbx, (%rdi) movq 88(%rsi), %r8 movq %rax, 8(%rdi) movq 80(%rsi), %r10 movq %r12, 16(%rdi) movq 72(%rsi), %r12 movq %rbp, 24(%rdi) movq 40(%rsi), %rax adcq %rdx, %rax movq 64(%rsi), %rdx movq %r13, 32(%rdi) movq 56(%rsi), %r13 movq 48(%rsi), %rbp adcq %r9, %rbp movq %rax, 40(%rdi) adcq %r11, %r13 adcq %r15, %rdx adcq %r14, %r12 adcq -16(%rsp), %r10 # 8-byte Folded Reload adcq -8(%rsp), %r8 # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax movq %rbp, %rsi subq (%rcx), %rsi movq %r13, %rbx sbbq 8(%rcx), %rbx movq %rdx, %r9 sbbq 16(%rcx), %r9 movq %r12, %r11 sbbq 24(%rcx), %r11 movq %r10, %r14 sbbq 32(%rcx), %r14 movq %r8, %r15 sbbq 40(%rcx), %r15 sbbq $0, %rax andl $1, %eax cmovneq %rbp, %rsi movq %rsi, 48(%rdi) testb %al, %al cmovneq %r13, %rbx movq %rbx, 56(%rdi) cmovneq %rdx, %r9 movq %r9, 64(%rdi) cmovneq %r12, %r11 movq %r11, 72(%rdi) cmovneq %r10, %r14 movq %r14, 80(%rdi) cmovneq %r8, %r15 movq %r15, 88(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end93: .size mcl_fpDbl_add6L, .Lfunc_end93-mcl_fpDbl_add6L .globl mcl_fpDbl_sub6L .align 16, 0x90 .type mcl_fpDbl_sub6L,@function mcl_fpDbl_sub6L: # @mcl_fpDbl_sub6L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 88(%rdx), %r9 movq 80(%rdx), %r10 movq 72(%rdx), %r14 movq 16(%rsi), %r8 movq (%rsi), %r15 movq 8(%rsi), %r11 xorl %eax, %eax subq (%rdx), %r15 sbbq 8(%rdx), %r11 sbbq 16(%rdx), %r8 movq 24(%rsi), %rbx sbbq 24(%rdx), %rbx movq 32(%rsi), %r12 sbbq 32(%rdx), %r12 movq 64(%rdx), %r13 movq %r15, (%rdi) movq 56(%rdx), %rbp movq %r11, 8(%rdi) movq 48(%rdx), %r15 movq 40(%rdx), %rdx movq %r8, 16(%rdi) movq 88(%rsi), %r8 movq %rbx, 24(%rdi) movq 40(%rsi), %rbx sbbq %rdx, %rbx movq 80(%rsi), %r11 movq %r12, 32(%rdi) movq 48(%rsi), %rdx sbbq %r15, %rdx movq 72(%rsi), %r15 movq %rbx, 40(%rdi) movq 64(%rsi), %r12 movq 56(%rsi), %rsi sbbq %rbp, %rsi sbbq %r13, %r12 sbbq %r14, %r15 sbbq %r10, %r11 sbbq %r9, %r8 movl $0, %ebp sbbq $0, %rbp andl $1, %ebp movq (%rcx), %r14 cmoveq %rax, %r14 testb %bpl, %bpl movq 16(%rcx), %r9 cmoveq %rax, %r9 movq 8(%rcx), %rbp cmoveq %rax, %rbp movq 40(%rcx), %r10 cmoveq %rax, %r10 movq 32(%rcx), %rbx cmoveq %rax, %rbx cmovneq 24(%rcx), %rax addq %rdx, %r14 movq %r14, 48(%rdi) adcq %rsi, %rbp movq %rbp, 56(%rdi) adcq %r12, %r9 movq %r9, 64(%rdi) adcq %r15, %rax movq %rax, 72(%rdi) adcq %r11, %rbx movq %rbx, 80(%rdi) adcq %r8, %r10 movq %r10, 88(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end94: .size mcl_fpDbl_sub6L, .Lfunc_end94-mcl_fpDbl_sub6L .globl mcl_fp_mulUnitPre7L .align 16, 0x90 .type mcl_fp_mulUnitPre7L,@function mcl_fp_mulUnitPre7L: # @mcl_fp_mulUnitPre7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rcx, %rax mulq 48(%rsi) movq %rdx, %r10 movq %rax, -8(%rsp) # 8-byte Spill movq %rcx, %rax mulq 40(%rsi) movq %rdx, %r11 movq %rax, -16(%rsp) # 8-byte Spill movq %rcx, %rax mulq 32(%rsi) movq %rdx, %r15 movq %rax, %r14 movq %rcx, %rax mulq 24(%rsi) movq %rdx, %r13 movq %rax, %r12 movq %rcx, %rax mulq 16(%rsi) movq %rdx, %rbx movq %rax, %rbp movq %rcx, %rax mulq 8(%rsi) movq %rdx, %r8 movq %rax, %r9 movq %rcx, %rax mulq (%rsi) movq %rax, (%rdi) addq %r9, %rdx movq %rdx, 8(%rdi) adcq %rbp, %r8 movq %r8, 16(%rdi) adcq %r12, %rbx movq %rbx, 24(%rdi) adcq %r14, %r13 movq %r13, 32(%rdi) adcq -16(%rsp), %r15 # 8-byte Folded Reload movq %r15, 40(%rdi) adcq -8(%rsp), %r11 # 8-byte Folded Reload movq %r11, 48(%rdi) adcq $0, %r10 movq %r10, 56(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end95: .size mcl_fp_mulUnitPre7L, .Lfunc_end95-mcl_fp_mulUnitPre7L .globl mcl_fpDbl_mulPre7L .align 16, 0x90 .type mcl_fpDbl_mulPre7L,@function mcl_fpDbl_mulPre7L: # @mcl_fpDbl_mulPre7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $24, %rsp movq %rdx, 8(%rsp) # 8-byte Spill movq %rsi, %r9 movq %rdi, 16(%rsp) # 8-byte Spill movq (%r9), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 8(%r9), %r10 movq %r10, -64(%rsp) # 8-byte Spill movq (%rdx), %rsi mulq %rsi movq %rdx, -32(%rsp) # 8-byte Spill movq 16(%r9), %r11 movq %r11, -72(%rsp) # 8-byte Spill movq 24(%r9), %rbx movq %rbx, -56(%rsp) # 8-byte Spill movq 32(%r9), %rbp movq %rbp, -24(%rsp) # 8-byte Spill movq 40(%r9), %rcx movq %rcx, -16(%rsp) # 8-byte Spill movq 48(%r9), %r14 movq %rax, (%rdi) movq %r14, %rax mulq %rsi movq %rdx, %rdi movq %rax, (%rsp) # 8-byte Spill movq %rcx, %rax mulq %rsi movq %rdx, %rcx movq %rax, -40(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rsi movq %rdx, %rbp movq %rax, %r15 movq %rbx, %rax mulq %rsi movq %rdx, %rbx movq %rax, %r8 movq %r11, %rax mulq %rsi movq %rdx, %r12 movq %rax, %r13 movq %r10, %rax mulq %rsi movq %rdx, %rsi movq %rax, %r10 addq -32(%rsp), %r10 # 8-byte Folded Reload adcq %r13, %rsi adcq %r8, %r12 adcq %r15, %rbx adcq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, -48(%rsp) # 8-byte Spill adcq (%rsp), %rcx # 8-byte Folded Reload movq %rcx, -40(%rsp) # 8-byte Spill adcq $0, %rdi movq %rdi, -32(%rsp) # 8-byte Spill movq 8(%rsp), %r11 # 8-byte Reload movq 8(%r11), %rcx movq %r14, %rax mulq %rcx movq %rdx, %r14 movq %rax, (%rsp) # 8-byte Spill movq -16(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %r8 movq -24(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r13 movq -56(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %r15 movq -72(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %rbp movq -64(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %rdi movq -8(%rsp), %rax # 8-byte Reload mulq %rcx addq %r10, %rax movq 16(%rsp), %r10 # 8-byte Reload movq %rax, 8(%r10) adcq %rsi, %rdi adcq %r12, %rbp adcq %rbx, %r15 adcq -48(%rsp), %r13 # 8-byte Folded Reload movq %r8, %rcx adcq -40(%rsp), %rcx # 8-byte Folded Reload movq (%rsp), %rax # 8-byte Reload adcq -32(%rsp), %rax # 8-byte Folded Reload sbbq %r8, %r8 andl $1, %r8d addq %rdx, %rdi adcq -64(%rsp), %rbp # 8-byte Folded Reload adcq -72(%rsp), %r15 # 8-byte Folded Reload adcq -56(%rsp), %r13 # 8-byte Folded Reload adcq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -88(%rsp) # 8-byte Spill adcq -16(%rsp), %rax # 8-byte Folded Reload movq %rax, (%rsp) # 8-byte Spill adcq %r14, %r8 movq 48(%r9), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 16(%r11), %rcx mulq %rcx movq %rax, -96(%rsp) # 8-byte Spill movq %rdx, -40(%rsp) # 8-byte Spill movq 40(%r9), %rax movq %rax, -24(%rsp) # 8-byte Spill mulq %rcx movq %rax, -16(%rsp) # 8-byte Spill movq %rdx, -72(%rsp) # 8-byte Spill movq 32(%r9), %rax movq %rax, -32(%rsp) # 8-byte Spill mulq %rcx movq %rax, %r12 movq %rdx, -80(%rsp) # 8-byte Spill movq 24(%r9), %rax movq %rax, -48(%rsp) # 8-byte Spill mulq %rcx movq %rax, %r14 movq %rdx, -104(%rsp) # 8-byte Spill movq 16(%r9), %rax movq %rax, -56(%rsp) # 8-byte Spill mulq %rcx movq %rax, %rbx movq %rdx, -112(%rsp) # 8-byte Spill movq (%r9), %rsi movq 8(%r9), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rcx movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r11 movq %rsi, %rax mulq %rcx addq %rdi, %rax movq %rax, 16(%r10) adcq %rbp, %r11 adcq %r15, %rbx adcq %r13, %r14 adcq -88(%rsp), %r12 # 8-byte Folded Reload movq -16(%rsp), %rdi # 8-byte Reload adcq (%rsp), %rdi # 8-byte Folded Reload movq -96(%rsp), %rax # 8-byte Reload adcq %r8, %rax sbbq %rcx, %rcx andl $1, %ecx addq %rdx, %r11 adcq -120(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %r14 # 8-byte Folded Reload adcq -104(%rsp), %r12 # 8-byte Folded Reload adcq -80(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -16(%rsp) # 8-byte Spill adcq -72(%rsp), %rax # 8-byte Folded Reload movq %rax, %rdi adcq -40(%rsp), %rcx # 8-byte Folded Reload movq 8(%rsp), %rax # 8-byte Reload movq 24(%rax), %rbp movq -8(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, (%rsp) # 8-byte Spill movq -24(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, -8(%rsp) # 8-byte Spill movq -32(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r13 movq -48(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %r15 movq -56(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %r8 movq -64(%rsp), %rax # 8-byte Reload mulq %rbp movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %r10 movq %rsi, %rax mulq %rbp addq %r11, %rax movq 16(%rsp), %rsi # 8-byte Reload movq %rax, 24(%rsi) adcq %rbx, %r10 adcq %r14, %r8 adcq %r12, %r15 adcq -16(%rsp), %r13 # 8-byte Folded Reload movq -8(%rsp), %rsi # 8-byte Reload adcq %rdi, %rsi movq (%rsp), %rax # 8-byte Reload adcq %rcx, %rax sbbq %rdi, %rdi andl $1, %edi addq %rdx, %r10 adcq -64(%rsp), %r8 # 8-byte Folded Reload adcq -56(%rsp), %r15 # 8-byte Folded Reload adcq -48(%rsp), %r13 # 8-byte Folded Reload adcq -32(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -8(%rsp) # 8-byte Spill adcq -24(%rsp), %rax # 8-byte Folded Reload movq %rax, (%rsp) # 8-byte Spill adcq -40(%rsp), %rdi # 8-byte Folded Reload movq 48(%r9), %rax movq %rax, -32(%rsp) # 8-byte Spill movq 8(%rsp), %rbx # 8-byte Reload movq 32(%rbx), %rcx mulq %rcx movq %rax, -16(%rsp) # 8-byte Spill movq %rdx, -56(%rsp) # 8-byte Spill movq 40(%r9), %rax movq %rax, -40(%rsp) # 8-byte Spill mulq %rcx movq %rax, -24(%rsp) # 8-byte Spill movq %rdx, -96(%rsp) # 8-byte Spill movq 32(%r9), %rax movq %rax, -48(%rsp) # 8-byte Spill mulq %rcx movq %rax, %r12 movq %rdx, -104(%rsp) # 8-byte Spill movq 24(%r9), %rax movq %rax, -64(%rsp) # 8-byte Spill mulq %rcx movq %rax, %rbp movq %rdx, -112(%rsp) # 8-byte Spill movq 16(%r9), %rax movq %rax, -72(%rsp) # 8-byte Spill mulq %rcx movq %rax, %r14 movq %rdx, -120(%rsp) # 8-byte Spill movq (%r9), %rsi movq %rsi, -80(%rsp) # 8-byte Spill movq 8(%r9), %rax movq %rax, -88(%rsp) # 8-byte Spill mulq %rcx movq %rdx, -128(%rsp) # 8-byte Spill movq %rax, %r11 movq %rsi, %rax mulq %rcx addq %r10, %rax movq 16(%rsp), %rcx # 8-byte Reload movq %rax, 32(%rcx) adcq %r8, %r11 adcq %r15, %r14 adcq %r13, %rbp adcq -8(%rsp), %r12 # 8-byte Folded Reload movq -24(%rsp), %rcx # 8-byte Reload adcq (%rsp), %rcx # 8-byte Folded Reload movq -16(%rsp), %rax # 8-byte Reload adcq %rdi, %rax sbbq %r13, %r13 andl $1, %r13d addq %rdx, %r11 adcq -128(%rsp), %r14 # 8-byte Folded Reload adcq -120(%rsp), %rbp # 8-byte Folded Reload adcq -112(%rsp), %r12 # 8-byte Folded Reload adcq -104(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -24(%rsp) # 8-byte Spill adcq -96(%rsp), %rax # 8-byte Folded Reload movq %rax, -16(%rsp) # 8-byte Spill adcq -56(%rsp), %r13 # 8-byte Folded Reload movq 40(%rbx), %rcx movq -32(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, (%rsp) # 8-byte Spill movq %rax, %rdi movq -40(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, %r10 movq -48(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r15 movq -64(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %rbx movq -72(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %rsi movq -88(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %r8 movq -80(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -96(%rsp) # 8-byte Spill addq %r11, %rax movq 16(%rsp), %rcx # 8-byte Reload movq %rax, 40(%rcx) adcq %r14, %r8 adcq %rbp, %rsi adcq %r12, %rbx adcq -24(%rsp), %r15 # 8-byte Folded Reload adcq -16(%rsp), %r10 # 8-byte Folded Reload adcq %r13, %rdi movq 8(%rsp), %rax # 8-byte Reload movq 48(%rax), %r11 sbbq %rcx, %rcx movq %r11, %rax mulq 48(%r9) movq %rdx, 8(%rsp) # 8-byte Spill movq %rax, -64(%rsp) # 8-byte Spill movq %r11, %rax mulq 40(%r9) movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %r11, %rax mulq 32(%r9) movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r13 movq %r11, %rax mulq 24(%r9) movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %rbp movq %r11, %rax mulq 16(%r9) movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, %r14 movq %r11, %rax mulq 8(%r9) movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %r12 movq %r11, %rax mulq (%r9) andl $1, %ecx addq -96(%rsp), %r8 # 8-byte Folded Reload adcq -72(%rsp), %rsi # 8-byte Folded Reload adcq -48(%rsp), %rbx # 8-byte Folded Reload adcq -40(%rsp), %r15 # 8-byte Folded Reload adcq -32(%rsp), %r10 # 8-byte Folded Reload adcq -8(%rsp), %rdi # 8-byte Folded Reload adcq (%rsp), %rcx # 8-byte Folded Reload addq %rax, %r8 movq 16(%rsp), %r9 # 8-byte Reload movq %r8, 48(%r9) adcq %r12, %rsi adcq %r14, %rbx adcq %rbp, %r15 adcq %r13, %r10 adcq -88(%rsp), %rdi # 8-byte Folded Reload adcq -64(%rsp), %rcx # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax addq %rdx, %rsi adcq -104(%rsp), %rbx # 8-byte Folded Reload movq %r9, %rdx movq %rsi, 56(%rdx) movq %rbx, 64(%rdx) adcq -80(%rsp), %r15 # 8-byte Folded Reload movq %r15, 72(%rdx) adcq -56(%rsp), %r10 # 8-byte Folded Reload movq %r10, 80(%rdx) adcq -24(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 88(%rdx) adcq -16(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 96(%rdx) adcq 8(%rsp), %rax # 8-byte Folded Reload movq %rax, 104(%rdx) addq $24, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end96: .size mcl_fpDbl_mulPre7L, .Lfunc_end96-mcl_fpDbl_mulPre7L .globl mcl_fpDbl_sqrPre7L .align 16, 0x90 .type mcl_fpDbl_sqrPre7L,@function mcl_fpDbl_sqrPre7L: # @mcl_fpDbl_sqrPre7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $16, %rsp movq %rdi, 8(%rsp) # 8-byte Spill movq 16(%rsi), %r11 movq %r11, -64(%rsp) # 8-byte Spill movq 24(%rsi), %r14 movq %r14, -48(%rsp) # 8-byte Spill movq 32(%rsi), %r9 movq %r9, -24(%rsp) # 8-byte Spill movq 40(%rsi), %r10 movq %r10, -16(%rsp) # 8-byte Spill movq 48(%rsi), %r8 movq (%rsi), %rbp movq 8(%rsi), %rbx movq %rbp, %rax mulq %rbp movq %rdx, %rcx movq %rax, (%rdi) movq %r8, %rax mulq %rbp movq %rdx, %r15 movq %rax, (%rsp) # 8-byte Spill movq %r10, %rax mulq %rbp movq %rdx, %rdi movq %rax, -32(%rsp) # 8-byte Spill movq %r9, %rax mulq %rbp movq %rdx, %r9 movq %rax, %r10 movq %r14, %rax mulq %rbp movq %rdx, %r13 movq %rax, %r14 movq %r11, %rax mulq %rbp movq %rdx, %r12 movq %rax, %r11 movq %rbx, %rax mulq %rbp movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, -56(%rsp) # 8-byte Spill addq %rax, %rcx adcq %rdx, %r11 adcq %r14, %r12 adcq %r10, %r13 adcq -32(%rsp), %r9 # 8-byte Folded Reload adcq (%rsp), %rdi # 8-byte Folded Reload movq %rdi, -40(%rsp) # 8-byte Spill adcq $0, %r15 movq %r15, -32(%rsp) # 8-byte Spill movq %r8, %rax mulq %rbx movq %rdx, (%rsp) # 8-byte Spill movq %rax, %rdi movq -16(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %r15 movq -24(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %r10 movq -48(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %r14 movq -64(%rsp), %rax # 8-byte Reload mulq %rbx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %rbp movq %rbx, %rax mulq %rbx movq %rax, %rbx addq -56(%rsp), %rcx # 8-byte Folded Reload movq 8(%rsp), %rax # 8-byte Reload movq %rcx, 8(%rax) adcq %r11, %rbx adcq %r12, %rbp adcq %r13, %r14 adcq %r9, %r10 adcq -40(%rsp), %r15 # 8-byte Folded Reload adcq -32(%rsp), %rdi # 8-byte Folded Reload sbbq %r8, %r8 andl $1, %r8d addq -8(%rsp), %rbx # 8-byte Folded Reload adcq %rdx, %rbp adcq -64(%rsp), %r14 # 8-byte Folded Reload adcq -48(%rsp), %r10 # 8-byte Folded Reload adcq -24(%rsp), %r15 # 8-byte Folded Reload adcq -16(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -72(%rsp) # 8-byte Spill adcq (%rsp), %r8 # 8-byte Folded Reload movq 48(%rsi), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 16(%rsi), %rdi mulq %rdi movq %rax, (%rsp) # 8-byte Spill movq %rdx, -40(%rsp) # 8-byte Spill movq 40(%rsi), %rax movq %rax, -24(%rsp) # 8-byte Spill mulq %rdi movq %rax, -88(%rsp) # 8-byte Spill movq %rdx, -56(%rsp) # 8-byte Spill movq 32(%rsi), %rax movq %rax, -32(%rsp) # 8-byte Spill mulq %rdi movq %rax, %r13 movq %rdx, -64(%rsp) # 8-byte Spill movq 24(%rsi), %rcx movq %rcx, %rax mulq %rdi movq %rax, %r9 movq %r9, -104(%rsp) # 8-byte Spill movq %rdx, -16(%rsp) # 8-byte Spill movq (%rsi), %r12 movq %r12, -48(%rsp) # 8-byte Spill movq 8(%rsi), %rax movq %rax, -80(%rsp) # 8-byte Spill mulq %rdi movq %rdx, -96(%rsp) # 8-byte Spill movq %rax, %r11 movq %r12, %rax mulq %rdi movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r12 movq %rdi, %rax mulq %rdi movq %rax, %rdi addq %rbx, %r12 movq 8(%rsp), %rax # 8-byte Reload movq %r12, 16(%rax) adcq %rbp, %r11 adcq %r14, %rdi adcq %r9, %r10 adcq %r15, %r13 movq -88(%rsp), %r14 # 8-byte Reload adcq -72(%rsp), %r14 # 8-byte Folded Reload movq (%rsp), %rax # 8-byte Reload adcq %r8, %rax sbbq %rbx, %rbx andl $1, %ebx addq -112(%rsp), %r11 # 8-byte Folded Reload adcq -96(%rsp), %rdi # 8-byte Folded Reload adcq %rdx, %r10 adcq -16(%rsp), %r13 # 8-byte Folded Reload adcq -64(%rsp), %r14 # 8-byte Folded Reload adcq -56(%rsp), %rax # 8-byte Folded Reload movq %rax, (%rsp) # 8-byte Spill adcq -40(%rsp), %rbx # 8-byte Folded Reload movq -8(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %r8 movq -24(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -8(%rsp) # 8-byte Spill movq -32(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %r9 movq -80(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, %r15 movq -48(%rsp), %rax # 8-byte Reload mulq %rcx movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, %rbp movq %rcx, %rax mulq %rcx movq %rax, %r12 movq %rdx, -96(%rsp) # 8-byte Spill addq %r11, %rbp movq 8(%rsp), %rax # 8-byte Reload movq %rbp, 24(%rax) adcq %rdi, %r15 adcq -104(%rsp), %r10 # 8-byte Folded Reload adcq %r13, %r12 movq %r9, %rcx adcq %r14, %rcx movq -8(%rsp), %rdi # 8-byte Reload adcq (%rsp), %rdi # 8-byte Folded Reload adcq %rbx, %r8 sbbq %r14, %r14 andl $1, %r14d movq (%rsi), %r9 movq 8(%rsi), %rbp movq 40(%rsi), %r11 movq %rbp, %rax mulq %r11 movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, -40(%rsp) # 8-byte Spill movq %r9, %rax mulq %r11 movq %rax, -48(%rsp) # 8-byte Spill movq %rdx, -32(%rsp) # 8-byte Spill movq 32(%rsi), %rbx movq %rbp, %rax mulq %rbx movq %rdx, (%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %r9, %rax mulq %rbx movq %rax, -120(%rsp) # 8-byte Spill movq %rdx, -104(%rsp) # 8-byte Spill addq -88(%rsp), %r15 # 8-byte Folded Reload adcq -80(%rsp), %r10 # 8-byte Folded Reload adcq -16(%rsp), %r12 # 8-byte Folded Reload adcq -96(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -96(%rsp) # 8-byte Spill adcq -72(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -8(%rsp) # 8-byte Spill adcq -64(%rsp), %r8 # 8-byte Folded Reload adcq -56(%rsp), %r14 # 8-byte Folded Reload movq 48(%rsi), %rax movq %rax, -56(%rsp) # 8-byte Spill mulq %rbx movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %rcx movq %r11, %rax mulq %rbx movq %rax, %rbp movq %rdx, -16(%rsp) # 8-byte Spill movq 24(%rsi), %rax movq %rax, -72(%rsp) # 8-byte Spill mulq %rbx movq %rax, %rdi movq %rdx, -88(%rsp) # 8-byte Spill movq 16(%rsi), %rax movq %rax, -80(%rsp) # 8-byte Spill mulq %rbx movq %rdx, -128(%rsp) # 8-byte Spill movq %rax, %r9 movq %rbx, %rax mulq %rbx movq %rax, %r13 addq -120(%rsp), %r15 # 8-byte Folded Reload movq 8(%rsp), %rax # 8-byte Reload movq %r15, 32(%rax) adcq -112(%rsp), %r10 # 8-byte Folded Reload adcq %r12, %r9 adcq -96(%rsp), %rdi # 8-byte Folded Reload adcq -8(%rsp), %r13 # 8-byte Folded Reload adcq %rbp, %r8 adcq %r14, %rcx sbbq %rbx, %rbx andl $1, %ebx addq -104(%rsp), %r10 # 8-byte Folded Reload adcq (%rsp), %r9 # 8-byte Folded Reload adcq -128(%rsp), %rdi # 8-byte Folded Reload adcq -88(%rsp), %r13 # 8-byte Folded Reload adcq %rdx, %r8 adcq -16(%rsp), %rcx # 8-byte Folded Reload adcq -64(%rsp), %rbx # 8-byte Folded Reload movq -56(%rsp), %rax # 8-byte Reload mulq %r11 movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, (%rsp) # 8-byte Spill movq -72(%rsp), %rax # 8-byte Reload mulq %r11 movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, %r15 movq -80(%rsp), %rax # 8-byte Reload mulq %r11 movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, %r14 movq %r11, %rax mulq %r11 movq %rax, %r12 addq -48(%rsp), %r10 # 8-byte Folded Reload movq 8(%rsp), %rax # 8-byte Reload movq %r10, 40(%rax) adcq -40(%rsp), %r9 # 8-byte Folded Reload adcq %rdi, %r14 adcq %r13, %r15 adcq %rbp, %r8 adcq %rcx, %r12 movq (%rsp), %rax # 8-byte Reload adcq %rbx, %rax sbbq %r11, %r11 andl $1, %r11d addq -32(%rsp), %r9 # 8-byte Folded Reload adcq -24(%rsp), %r14 # 8-byte Folded Reload adcq -64(%rsp), %r15 # 8-byte Folded Reload adcq -56(%rsp), %r8 # 8-byte Folded Reload movq %r8, -32(%rsp) # 8-byte Spill adcq -16(%rsp), %r12 # 8-byte Folded Reload adcq %rdx, %rax movq %rax, (%rsp) # 8-byte Spill adcq -8(%rsp), %r11 # 8-byte Folded Reload movq 48(%rsi), %rcx movq %rcx, %rax mulq 40(%rsi) movq %rdx, -8(%rsp) # 8-byte Spill movq %rax, %r8 movq %rcx, %rax mulq 32(%rsi) movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %rbx movq %rcx, %rax mulq 24(%rsi) movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %rbp movq %rcx, %rax mulq 16(%rsi) movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %r10 movq %rcx, %rax mulq 8(%rsi) movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, %rdi movq %rcx, %rax mulq (%rsi) movq %rdx, %r13 movq %rax, %rsi movq %rcx, %rax mulq %rcx addq %r9, %rsi movq 8(%rsp), %r9 # 8-byte Reload movq %rsi, 48(%r9) adcq %r14, %rdi adcq %r15, %r10 adcq -32(%rsp), %rbp # 8-byte Folded Reload adcq %r12, %rbx adcq (%rsp), %r8 # 8-byte Folded Reload adcq %r11, %rax sbbq %rcx, %rcx andl $1, %ecx addq %r13, %rdi adcq -48(%rsp), %r10 # 8-byte Folded Reload movq %r9, %rsi movq %rdi, 56(%rsi) movq %r10, 64(%rsi) adcq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 72(%rsi) adcq -24(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 80(%rsi) adcq -16(%rsp), %r8 # 8-byte Folded Reload movq %r8, 88(%rsi) adcq -8(%rsp), %rax # 8-byte Folded Reload movq %rax, 96(%rsi) adcq %rdx, %rcx movq %rcx, 104(%rsi) addq $16, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end97: .size mcl_fpDbl_sqrPre7L, .Lfunc_end97-mcl_fpDbl_sqrPre7L .globl mcl_fp_mont7L .align 16, 0x90 .type mcl_fp_mont7L,@function mcl_fp_mont7L: # @mcl_fp_mont7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $96, %rsp movq %rdx, 24(%rsp) # 8-byte Spill movq %rdi, -96(%rsp) # 8-byte Spill movq 48(%rsi), %rax movq %rax, 16(%rsp) # 8-byte Spill movq (%rdx), %rbx mulq %rbx movq %rax, 88(%rsp) # 8-byte Spill movq %rdx, %r15 movq 40(%rsi), %rax movq %rax, 8(%rsp) # 8-byte Spill mulq %rbx movq %rax, 80(%rsp) # 8-byte Spill movq %rdx, %r12 movq 32(%rsi), %rax movq %rax, -32(%rsp) # 8-byte Spill movq 24(%rsi), %r9 movq %r9, -24(%rsp) # 8-byte Spill movq 16(%rsi), %r10 movq %r10, -16(%rsp) # 8-byte Spill movq (%rsi), %r13 movq %r13, (%rsp) # 8-byte Spill movq 8(%rsi), %rsi movq %rsi, -8(%rsp) # 8-byte Spill mulq %rbx movq %rdx, %r14 movq %rax, %r8 movq %r9, %rax mulq %rbx movq %rdx, %rdi movq %rax, %r9 movq %r10, %rax mulq %rbx movq %rdx, %rbp movq %rax, %r10 movq %rsi, %rax mulq %rbx movq %rdx, %rsi movq %rax, %r11 movq %r13, %rax mulq %rbx movq %rax, -112(%rsp) # 8-byte Spill addq %r11, %rdx movq %rdx, -104(%rsp) # 8-byte Spill adcq %r10, %rsi movq %rsi, -88(%rsp) # 8-byte Spill adcq %r9, %rbp movq %rbp, -80(%rsp) # 8-byte Spill adcq %r8, %rdi movq %rdi, -72(%rsp) # 8-byte Spill adcq 80(%rsp), %r14 # 8-byte Folded Reload movq %r14, -64(%rsp) # 8-byte Spill adcq 88(%rsp), %r12 # 8-byte Folded Reload movq %r12, -48(%rsp) # 8-byte Spill adcq $0, %r15 movq %r15, -40(%rsp) # 8-byte Spill movq -8(%rcx), %rdx movq %rdx, 32(%rsp) # 8-byte Spill movq %rax, %rdi imulq %rdx, %rdi movq (%rcx), %r12 movq %r12, 40(%rsp) # 8-byte Spill movq 48(%rcx), %rdx movq %rdx, 64(%rsp) # 8-byte Spill movq 40(%rcx), %r9 movq %r9, 88(%rsp) # 8-byte Spill movq 32(%rcx), %rbx movq %rbx, 80(%rsp) # 8-byte Spill movq 24(%rcx), %rsi movq %rsi, 72(%rsp) # 8-byte Spill movq 16(%rcx), %rbp movq %rbp, 48(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, 56(%rsp) # 8-byte Spill movq %rdi, %rax mulq %rdx movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rdi, %rax mulq %r9 movq %rdx, %r14 movq %rax, -128(%rsp) # 8-byte Spill movq %rdi, %rax mulq %rbx movq %rdx, %r11 movq %rax, %r15 movq %rdi, %rax mulq %rsi movq %rdx, %rbx movq %rax, %r10 movq %rdi, %rax mulq %rbp movq %rdx, %r8 movq %rax, %r13 movq %rdi, %rax mulq %rcx movq %rdx, %rbp movq %rax, %r9 movq %rdi, %rax mulq %r12 movq %rdx, %r12 addq %r9, %r12 adcq %r13, %rbp adcq %r10, %r8 adcq %r15, %rbx adcq -128(%rsp), %r11 # 8-byte Folded Reload adcq -120(%rsp), %r14 # 8-byte Folded Reload movq -56(%rsp), %rcx # 8-byte Reload adcq $0, %rcx addq -112(%rsp), %rax # 8-byte Folded Reload adcq -104(%rsp), %r12 # 8-byte Folded Reload adcq -88(%rsp), %rbp # 8-byte Folded Reload adcq -80(%rsp), %r8 # 8-byte Folded Reload adcq -72(%rsp), %rbx # 8-byte Folded Reload adcq -64(%rsp), %r11 # 8-byte Folded Reload adcq -48(%rsp), %r14 # 8-byte Folded Reload adcq -40(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -56(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -72(%rsp) # 8-byte Spill movq 24(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdi movq %rdi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %r13 movq %rdi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %r10 movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill movq %rdx, %rdi addq %r10, %rdi adcq %r13, %r15 adcq -112(%rsp), %r9 # 8-byte Folded Reload movq %rcx, %rdx adcq -104(%rsp), %rdx # 8-byte Folded Reload adcq -88(%rsp), %rsi # 8-byte Folded Reload movq -48(%rsp), %rax # 8-byte Reload adcq -80(%rsp), %rax # 8-byte Folded Reload movq -40(%rsp), %rcx # 8-byte Reload adcq $0, %rcx movq -64(%rsp), %r10 # 8-byte Reload addq %r12, %r10 movq %r10, -64(%rsp) # 8-byte Spill adcq %rbp, %rdi adcq %r8, %r15 adcq %rbx, %r9 adcq %r11, %rdx movq %rdx, -88(%rsp) # 8-byte Spill adcq %r14, %rsi movq %rsi, -80(%rsp) # 8-byte Spill adcq -56(%rsp), %rax # 8-byte Folded Reload movq %rax, -48(%rsp) # 8-byte Spill adcq -72(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -40(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -56(%rsp) # 8-byte Spill movq %r10, %rbp imulq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq 88(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq 80(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %rbx movq %rbp, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r10 movq %rbp, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r14 movq %rbp, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r11 addq %r14, %r11 adcq %r10, %rsi adcq %rbx, %rcx adcq -120(%rsp), %r13 # 8-byte Folded Reload adcq -112(%rsp), %r12 # 8-byte Folded Reload adcq -104(%rsp), %r8 # 8-byte Folded Reload movq -72(%rsp), %rbp # 8-byte Reload adcq $0, %rbp addq -64(%rsp), %rax # 8-byte Folded Reload adcq %rdi, %r11 adcq %r15, %rsi adcq %r9, %rcx adcq -88(%rsp), %r13 # 8-byte Folded Reload adcq -80(%rsp), %r12 # 8-byte Folded Reload adcq -48(%rsp), %r8 # 8-byte Folded Reload movq %r8, -80(%rsp) # 8-byte Spill adcq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, -72(%rsp) # 8-byte Spill movq -56(%rsp), %rbp # 8-byte Reload adcq $0, %rbp movq 24(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdi movq %rdi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %r9 movq %rdi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r14 movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rax, %r10 movq %rdx, %r8 addq %r14, %r8 adcq %r9, %rbx adcq -120(%rsp), %r15 # 8-byte Folded Reload movq -64(%rsp), %r9 # 8-byte Reload adcq -112(%rsp), %r9 # 8-byte Folded Reload movq -56(%rsp), %rdi # 8-byte Reload adcq -104(%rsp), %rdi # 8-byte Folded Reload movq -48(%rsp), %rdx # 8-byte Reload adcq -88(%rsp), %rdx # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r11, %r10 adcq %rsi, %r8 adcq %rcx, %rbx adcq %r13, %r15 adcq %r12, %r9 movq %r9, -64(%rsp) # 8-byte Spill adcq -80(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -56(%rsp) # 8-byte Spill adcq -72(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill adcq %rbp, %rax movq %rax, -40(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -88(%rsp) # 8-byte Spill movq %r10, %rbp imulq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq 88(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq 80(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %rcx movq %rbp, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r13 movq %rbp, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r12 movq %rbp, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r14 addq %r12, %r14 adcq %r13, %rsi adcq %rcx, %rdi adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq -104(%rsp), %rdx # 8-byte Folded Reload movq -72(%rsp), %rcx # 8-byte Reload adcq $0, %rcx addq %r10, %rax adcq %r8, %r14 adcq %rbx, %rsi adcq %r15, %rdi adcq -64(%rsp), %r9 # 8-byte Folded Reload adcq -56(%rsp), %r11 # 8-byte Folded Reload adcq -48(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill adcq -40(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -72(%rsp) # 8-byte Spill adcq $0, -88(%rsp) # 8-byte Folded Spill movq 24(%rsp), %rax # 8-byte Reload movq 24(%rax), %rcx movq %rcx, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, -64(%rsp) # 8-byte Spill movq %rcx, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r8 movq %rcx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r12 movq %rcx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rax, %r10 movq %rdx, %r13 addq %r12, %r13 adcq %r8, %rbp adcq -120(%rsp), %rbx # 8-byte Folded Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload movq -56(%rsp), %rdx # 8-byte Reload adcq -104(%rsp), %rdx # 8-byte Folded Reload movq -48(%rsp), %rcx # 8-byte Reload adcq -64(%rsp), %rcx # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r14, %r10 adcq %rsi, %r13 adcq %rdi, %rbp adcq %r9, %rbx adcq %r11, %r15 adcq -80(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill adcq -72(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -48(%rsp) # 8-byte Spill adcq -88(%rsp), %rax # 8-byte Folded Reload movq %rax, -40(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -64(%rsp) # 8-byte Spill movq %r10, %rsi imulq 32(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rsi, %rax mulq 88(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq 80(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r9 movq %rsi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r11 movq %rsi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r14 addq %r11, %r14 adcq %r9, %r8 adcq -120(%rsp), %rcx # 8-byte Folded Reload adcq -112(%rsp), %rdi # 8-byte Folded Reload adcq -104(%rsp), %r12 # 8-byte Folded Reload movq -80(%rsp), %rsi # 8-byte Reload adcq -88(%rsp), %rsi # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r10, %rax adcq %r13, %r14 adcq %rbp, %r8 adcq %rbx, %rcx adcq %r15, %rdi adcq -56(%rsp), %r12 # 8-byte Folded Reload adcq -48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -80(%rsp) # 8-byte Spill adcq -40(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq -64(%rsp), %r11 # 8-byte Reload adcq $0, %r11 movq 24(%rsp), %rax # 8-byte Reload movq 32(%rax), %rbp movq %rbp, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rbp, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r9 movq %rbp, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r15 movq %rbp, %rax mulq (%rsp) # 8-byte Folded Reload movq %rax, %rsi movq %rdx, %r10 addq %r15, %r10 adcq %r9, %r13 adcq -120(%rsp), %rbx # 8-byte Folded Reload movq -64(%rsp), %r15 # 8-byte Reload adcq -112(%rsp), %r15 # 8-byte Folded Reload movq -56(%rsp), %rbp # 8-byte Reload adcq -104(%rsp), %rbp # 8-byte Folded Reload movq -48(%rsp), %rdx # 8-byte Reload adcq -88(%rsp), %rdx # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload adcq $0, %rax movq %rsi, %r9 addq %r14, %r9 adcq %r8, %r10 adcq %rcx, %r13 adcq %rdi, %rbx adcq %r12, %r15 movq %r15, -64(%rsp) # 8-byte Spill adcq -80(%rsp), %rbp # 8-byte Folded Reload movq %rbp, -56(%rsp) # 8-byte Spill adcq -72(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill adcq %r11, %rax movq %rax, -40(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -88(%rsp) # 8-byte Spill movq %r9, %rsi movq %r9, %r11 imulq 32(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq 88(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq 80(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %rdi movq %rsi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r12 movq %rsi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r15 movq %rsi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r14 addq %r15, %r14 adcq %r12, %rcx adcq %rdi, %rbp adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq -112(%rsp), %r8 # 8-byte Folded Reload movq -80(%rsp), %rsi # 8-byte Reload adcq -104(%rsp), %rsi # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r11, %rax adcq %r10, %r14 adcq %r13, %rcx adcq %rbx, %rbp adcq -64(%rsp), %r9 # 8-byte Folded Reload adcq -56(%rsp), %r8 # 8-byte Folded Reload adcq -48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -80(%rsp) # 8-byte Spill adcq -40(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill adcq $0, -88(%rsp) # 8-byte Folded Spill movq 24(%rsp), %rax # 8-byte Reload movq 40(%rax), %rdi movq %rdi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, -64(%rsp) # 8-byte Spill movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, %r10 movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r15 movq %rdi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r13 movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rax, %rdi movq %rdx, %r11 addq %r13, %r11 adcq %r15, %rsi adcq %r10, %rbx adcq -112(%rsp), %r12 # 8-byte Folded Reload movq -56(%rsp), %r10 # 8-byte Reload adcq -104(%rsp), %r10 # 8-byte Folded Reload movq -48(%rsp), %rdx # 8-byte Reload adcq -64(%rsp), %rdx # 8-byte Folded Reload movq -40(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %r14, %rdi adcq %rcx, %r11 adcq %rbp, %rsi adcq %r9, %rbx adcq %r8, %r12 adcq -80(%rsp), %r10 # 8-byte Folded Reload movq %r10, -56(%rsp) # 8-byte Spill adcq -72(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill adcq -88(%rsp), %rax # 8-byte Folded Reload movq %rax, -40(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, -80(%rsp) # 8-byte Spill movq %rdi, %rbp imulq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rbp, %rax mulq 88(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbp, %rax mulq 80(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -112(%rsp) # 8-byte Spill movq %rbp, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, -120(%rsp) # 8-byte Spill movq %rbp, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %r8 movq %rbp, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r9 movq %rbp, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r15 addq %r9, %r15 adcq %r8, %r13 adcq -120(%rsp), %r10 # 8-byte Folded Reload adcq -112(%rsp), %rcx # 8-byte Folded Reload adcq -104(%rsp), %r14 # 8-byte Folded Reload movq -72(%rsp), %rdx # 8-byte Reload adcq -88(%rsp), %rdx # 8-byte Folded Reload movq -64(%rsp), %r8 # 8-byte Reload adcq $0, %r8 addq %rdi, %rax adcq %r11, %r15 adcq %rsi, %r13 adcq %rbx, %r10 adcq %r12, %rcx adcq -56(%rsp), %r14 # 8-byte Folded Reload movq %r14, -56(%rsp) # 8-byte Spill adcq -48(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill adcq -40(%rsp), %r8 # 8-byte Folded Reload movq %r8, -64(%rsp) # 8-byte Spill movq -80(%rsp), %rsi # 8-byte Reload adcq $0, %rsi movq 24(%rsp), %rax # 8-byte Reload movq 48(%rax), %rdi movq %rdi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, 16(%rsp) # 8-byte Spill movq %rax, -40(%rsp) # 8-byte Spill movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, 24(%rsp) # 8-byte Spill movq %rax, 8(%rsp) # 8-byte Spill movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -32(%rsp) # 8-byte Spill movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r12 movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %rbx movq %rdi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %rbp movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload addq %rbp, %rdx movq %rdx, %rbp adcq %rbx, %r9 adcq %r12, %r14 movq %r8, %rdi adcq -32(%rsp), %rdi # 8-byte Folded Reload adcq 8(%rsp), %r11 # 8-byte Folded Reload movq 24(%rsp), %rbx # 8-byte Reload adcq -40(%rsp), %rbx # 8-byte Folded Reload movq 16(%rsp), %r8 # 8-byte Reload adcq $0, %r8 addq %r15, %rax movq %rax, -32(%rsp) # 8-byte Spill adcq %r13, %rbp movq %rbp, 8(%rsp) # 8-byte Spill adcq %r10, %r9 movq %r9, (%rsp) # 8-byte Spill adcq %rcx, %r14 movq %r14, -8(%rsp) # 8-byte Spill adcq -56(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -16(%rsp) # 8-byte Spill adcq -72(%rsp), %r11 # 8-byte Folded Reload movq %r11, -24(%rsp) # 8-byte Spill adcq -64(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 24(%rsp) # 8-byte Spill adcq %rsi, %r8 movq %r8, 16(%rsp) # 8-byte Spill sbbq %rcx, %rcx movq 32(%rsp), %r10 # 8-byte Reload imulq %rax, %r10 andl $1, %ecx movq %r10, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, 32(%rsp) # 8-byte Spill movq %r10, %rax mulq 88(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -40(%rsp) # 8-byte Spill movq %r10, %rax mulq 80(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, -48(%rsp) # 8-byte Spill movq %r10, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, -56(%rsp) # 8-byte Spill movq %r10, %rax movq 48(%rsp), %r13 # 8-byte Reload mulq %r13 movq %rdx, %rbp movq %rax, %r12 movq %r10, %rax movq 40(%rsp), %r15 # 8-byte Reload mulq %r15 movq %rdx, %r11 movq %rax, %r8 movq %r10, %rax movq 56(%rsp), %r14 # 8-byte Reload mulq %r14 addq %r11, %rax adcq %r12, %rdx adcq -56(%rsp), %rbp # 8-byte Folded Reload adcq -48(%rsp), %rsi # 8-byte Folded Reload adcq -40(%rsp), %rdi # 8-byte Folded Reload adcq 32(%rsp), %r9 # 8-byte Folded Reload adcq $0, %rbx addq -32(%rsp), %r8 # 8-byte Folded Reload adcq 8(%rsp), %rax # 8-byte Folded Reload adcq (%rsp), %rdx # 8-byte Folded Reload adcq -8(%rsp), %rbp # 8-byte Folded Reload adcq -16(%rsp), %rsi # 8-byte Folded Reload adcq -24(%rsp), %rdi # 8-byte Folded Reload adcq 24(%rsp), %r9 # 8-byte Folded Reload adcq 16(%rsp), %rbx # 8-byte Folded Reload adcq $0, %rcx movq %rax, %r8 subq %r15, %r8 movq %rdx, %r10 sbbq %r14, %r10 movq %rbp, %r11 sbbq %r13, %r11 movq %rsi, %r14 sbbq 72(%rsp), %r14 # 8-byte Folded Reload movq %rdi, %r15 sbbq 80(%rsp), %r15 # 8-byte Folded Reload movq %r9, %r12 sbbq 88(%rsp), %r12 # 8-byte Folded Reload movq %rbx, %r13 sbbq 64(%rsp), %r13 # 8-byte Folded Reload sbbq $0, %rcx andl $1, %ecx cmovneq %rbx, %r13 testb %cl, %cl cmovneq %rax, %r8 movq -96(%rsp), %rax # 8-byte Reload movq %r8, (%rax) cmovneq %rdx, %r10 movq %r10, 8(%rax) cmovneq %rbp, %r11 movq %r11, 16(%rax) cmovneq %rsi, %r14 movq %r14, 24(%rax) cmovneq %rdi, %r15 movq %r15, 32(%rax) cmovneq %r9, %r12 movq %r12, 40(%rax) movq %r13, 48(%rax) addq $96, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end98: .size mcl_fp_mont7L, .Lfunc_end98-mcl_fp_mont7L .globl mcl_fp_montNF7L .align 16, 0x90 .type mcl_fp_montNF7L,@function mcl_fp_montNF7L: # @mcl_fp_montNF7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $80, %rsp movq %rdx, 16(%rsp) # 8-byte Spill movq %rdi, -96(%rsp) # 8-byte Spill movq 48(%rsi), %rax movq %rax, 8(%rsp) # 8-byte Spill movq (%rdx), %rbp mulq %rbp movq %rax, 72(%rsp) # 8-byte Spill movq %rdx, %r9 movq 40(%rsi), %rax movq %rax, (%rsp) # 8-byte Spill mulq %rbp movq %rax, 64(%rsp) # 8-byte Spill movq %rdx, %r11 movq 32(%rsi), %rax movq %rax, -48(%rsp) # 8-byte Spill movq 24(%rsi), %r8 movq %r8, -40(%rsp) # 8-byte Spill movq 16(%rsi), %rbx movq %rbx, -32(%rsp) # 8-byte Spill movq (%rsi), %r10 movq %r10, -16(%rsp) # 8-byte Spill movq 8(%rsi), %rsi movq %rsi, -24(%rsp) # 8-byte Spill mulq %rbp movq %rdx, %rdi movq %rax, 56(%rsp) # 8-byte Spill movq %r8, %rax mulq %rbp movq %rdx, %r14 movq %rax, %r15 movq %rbx, %rax mulq %rbp movq %rdx, %rbx movq %rax, %r13 movq %rsi, %rax mulq %rbp movq %rdx, %rsi movq %rax, %r12 movq %r10, %rax mulq %rbp movq %rdx, %r8 addq %r12, %r8 adcq %r13, %rsi movq %rsi, -104(%rsp) # 8-byte Spill adcq %r15, %rbx movq %rbx, -88(%rsp) # 8-byte Spill adcq 56(%rsp), %r14 # 8-byte Folded Reload movq %r14, %r12 adcq 64(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -80(%rsp) # 8-byte Spill adcq 72(%rsp), %r11 # 8-byte Folded Reload movq %r11, -56(%rsp) # 8-byte Spill adcq $0, %r9 movq %r9, -64(%rsp) # 8-byte Spill movq -8(%rcx), %rdx movq %rdx, 24(%rsp) # 8-byte Spill movq %rax, %r9 movq %rax, %r14 imulq %rdx, %r9 movq (%rcx), %r11 movq %r11, 32(%rsp) # 8-byte Spill movq 48(%rcx), %rdx movq %rdx, 72(%rsp) # 8-byte Spill movq 40(%rcx), %r10 movq %r10, 64(%rsp) # 8-byte Spill movq 32(%rcx), %rbp movq %rbp, 56(%rsp) # 8-byte Spill movq 24(%rcx), %rbx movq %rbx, 48(%rsp) # 8-byte Spill movq 16(%rcx), %rdi movq %rdi, 40(%rsp) # 8-byte Spill movq 8(%rcx), %rsi movq %rsi, -8(%rsp) # 8-byte Spill movq %r9, %rax mulq %rdx movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq %r9, %rax mulq %r10 movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r15 movq %r9, %rax mulq %rbp movq %rdx, -128(%rsp) # 8-byte Spill movq %rax, %r13 movq %r9, %rax mulq %rbx movq %rdx, %rbx movq %rax, %rbp movq %r9, %rax mulq %rdi movq %rdx, %rcx movq %rax, %rdi movq %r9, %rax mulq %rsi movq %rdx, %r10 movq %rax, %rsi movq %r9, %rax mulq %r11 addq %r14, %rax adcq %r8, %rsi adcq -104(%rsp), %rdi # 8-byte Folded Reload adcq -88(%rsp), %rbp # 8-byte Folded Reload adcq %r12, %r13 adcq -80(%rsp), %r15 # 8-byte Folded Reload movq -72(%rsp), %r8 # 8-byte Reload adcq -56(%rsp), %r8 # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rdx, %rsi adcq %r10, %rdi adcq %rcx, %rbp adcq %rbx, %r13 adcq -128(%rsp), %r15 # 8-byte Folded Reload adcq -120(%rsp), %r8 # 8-byte Folded Reload movq %r8, -72(%rsp) # 8-byte Spill adcq -112(%rsp), %rax # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 8(%rax), %rbx movq %rbx, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rbx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rbx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, -112(%rsp) # 8-byte Spill movq %rbx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -120(%rsp) # 8-byte Spill movq %rbx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -128(%rsp) # 8-byte Spill movq %rbx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r14 movq %rbx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rax, %r10 movq %rdx, %r12 addq %r14, %r12 adcq -128(%rsp), %rcx # 8-byte Folded Reload adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq -112(%rsp), %r8 # 8-byte Folded Reload adcq -104(%rsp), %r11 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq -88(%rsp), %rdx # 8-byte Folded Reload movq -56(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rsi, %r10 adcq %rdi, %r12 adcq %rbp, %rcx adcq %r13, %r9 adcq %r15, %r8 adcq -72(%rsp), %r11 # 8-byte Folded Reload adcq -64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, -56(%rsp) # 8-byte Spill movq %r10, %rbx imulq 24(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, -64(%rsp) # 8-byte Spill movq %rbx, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq %rbx, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r14 movq %rbx, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -128(%rsp) # 8-byte Spill movq %rax, %rdi movq %rbx, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %rbp movq %rbx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %rsi movq %rbx, %rax mulq 32(%rsp) # 8-byte Folded Reload addq %r10, %rax adcq %r12, %rsi adcq %rcx, %rbp adcq %r9, %rdi adcq %r8, %r14 movq -72(%rsp), %rcx # 8-byte Reload adcq %r11, %rcx movq -64(%rsp), %r8 # 8-byte Reload adcq -80(%rsp), %r8 # 8-byte Folded Reload movq -56(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rdx, %rsi adcq %r13, %rbp adcq %r15, %rdi movq %rdi, -88(%rsp) # 8-byte Spill adcq -128(%rsp), %r14 # 8-byte Folded Reload movq %r14, -80(%rsp) # 8-byte Spill adcq -120(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -72(%rsp) # 8-byte Spill adcq -112(%rsp), %r8 # 8-byte Folded Reload movq %r8, -64(%rsp) # 8-byte Spill adcq -104(%rsp), %rax # 8-byte Folded Reload movq %rax, -56(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 16(%rax), %rcx movq %rcx, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, -104(%rsp) # 8-byte Spill movq %rcx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r12 movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r9 movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %r11 movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rax, %r13 movq %rdx, %rcx addq %r11, %rcx adcq %r9, %r15 adcq %r12, %rbx adcq -120(%rsp), %rdi # 8-byte Folded Reload adcq -112(%rsp), %r14 # 8-byte Folded Reload adcq -104(%rsp), %r10 # 8-byte Folded Reload adcq $0, %r8 addq %rsi, %r13 adcq %rbp, %rcx adcq -88(%rsp), %r15 # 8-byte Folded Reload adcq -80(%rsp), %rbx # 8-byte Folded Reload adcq -72(%rsp), %rdi # 8-byte Folded Reload adcq -64(%rsp), %r14 # 8-byte Folded Reload adcq -56(%rsp), %r10 # 8-byte Folded Reload adcq $0, %r8 movq %r13, %r9 imulq 24(%rsp), %r9 # 8-byte Folded Reload movq %r9, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -64(%rsp) # 8-byte Spill movq %r9, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq %r9, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, %rbp movq %r9, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, %r11 movq %r9, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r12 movq %r9, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %rsi movq %r9, %rax mulq 32(%rsp) # 8-byte Folded Reload addq %r13, %rax adcq %rcx, %rsi adcq %r15, %r12 adcq %rbx, %r11 adcq %rdi, %rbp movq -72(%rsp), %rcx # 8-byte Reload adcq %r14, %rcx movq -64(%rsp), %rax # 8-byte Reload adcq %r10, %rax adcq $0, %r8 addq %rdx, %rsi adcq -120(%rsp), %r12 # 8-byte Folded Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload adcq -104(%rsp), %rbp # 8-byte Folded Reload adcq -88(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -72(%rsp) # 8-byte Spill adcq -80(%rsp), %rax # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill adcq -56(%rsp), %r8 # 8-byte Folded Reload movq 16(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdi movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -120(%rsp) # 8-byte Spill movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rbx movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r9 movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rax, %r14 movq %rdx, %rdi addq %r9, %rdi adcq %rbx, %rcx adcq -120(%rsp), %r10 # 8-byte Folded Reload adcq -112(%rsp), %r13 # 8-byte Folded Reload adcq -104(%rsp), %r15 # 8-byte Folded Reload movq -88(%rsp), %rdx # 8-byte Reload adcq -80(%rsp), %rdx # 8-byte Folded Reload movq -56(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rsi, %r14 adcq %r12, %rdi adcq %r11, %rcx adcq %rbp, %r10 adcq -72(%rsp), %r13 # 8-byte Folded Reload adcq -64(%rsp), %r15 # 8-byte Folded Reload adcq %r8, %rdx movq %rdx, -88(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, -56(%rsp) # 8-byte Spill movq %r14, %rsi imulq 24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq %rsi, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rsi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r12 movq %rsi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r8 movq %rsi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %rbp movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %rbx movq %rsi, %rax mulq 32(%rsp) # 8-byte Folded Reload addq %r14, %rax adcq %rdi, %rbx adcq %rcx, %rbp adcq %r10, %r8 adcq %r13, %r12 movq -80(%rsp), %rsi # 8-byte Reload adcq %r15, %rsi movq -72(%rsp), %rcx # 8-byte Reload adcq -88(%rsp), %rcx # 8-byte Folded Reload movq -56(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rdx, %rbx adcq %r9, %rbp adcq %r11, %r8 adcq -120(%rsp), %r12 # 8-byte Folded Reload adcq -112(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -80(%rsp) # 8-byte Spill adcq -104(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -72(%rsp) # 8-byte Spill adcq -64(%rsp), %rax # 8-byte Folded Reload movq %rax, -56(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 32(%rax), %rsi movq %rsi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -104(%rsp) # 8-byte Spill movq %rsi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rsi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -120(%rsp) # 8-byte Spill movq %rsi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -128(%rsp) # 8-byte Spill movq %rsi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r14 movq %rsi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r15 movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rax, %r11 movq %rdx, %r10 addq %r15, %r10 adcq %r14, %rdi adcq -128(%rsp), %rcx # 8-byte Folded Reload adcq -120(%rsp), %r9 # 8-byte Folded Reload adcq -112(%rsp), %r13 # 8-byte Folded Reload movq -88(%rsp), %rdx # 8-byte Reload adcq -104(%rsp), %rdx # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rbx, %r11 adcq %rbp, %r10 adcq %r8, %rdi adcq %r12, %rcx adcq -80(%rsp), %r9 # 8-byte Folded Reload adcq -72(%rsp), %r13 # 8-byte Folded Reload adcq -56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, -64(%rsp) # 8-byte Spill movq %r11, %rsi imulq 24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -56(%rsp) # 8-byte Spill movq %rsi, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq %rsi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r14 movq %rsi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %r12 movq %rsi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %rbp movq %rsi, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %rbx movq %rsi, %rax mulq 32(%rsp) # 8-byte Folded Reload addq %r11, %rax adcq %r10, %rbx adcq %rdi, %rbp adcq %rcx, %r12 adcq %r9, %r14 movq -72(%rsp), %rdi # 8-byte Reload adcq %r13, %rdi movq -56(%rsp), %rcx # 8-byte Reload adcq -88(%rsp), %rcx # 8-byte Folded Reload movq -64(%rsp), %rax # 8-byte Reload adcq $0, %rax addq %rdx, %rbx adcq %r8, %rbp adcq %r15, %r12 adcq -120(%rsp), %r14 # 8-byte Folded Reload movq %r14, -88(%rsp) # 8-byte Spill adcq -112(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -72(%rsp) # 8-byte Spill adcq -104(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -56(%rsp) # 8-byte Spill adcq -80(%rsp), %rax # 8-byte Folded Reload movq %rax, -64(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 40(%rax), %rcx movq %rcx, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, -104(%rsp) # 8-byte Spill movq %rcx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, -128(%rsp) # 8-byte Spill movq %rcx, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r15 movq %rcx, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r14 movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rax, %r11 movq %rdx, %r10 addq %r14, %r10 adcq %r15, %r8 adcq -128(%rsp), %rdi # 8-byte Folded Reload adcq -120(%rsp), %rsi # 8-byte Folded Reload adcq -112(%rsp), %r13 # 8-byte Folded Reload movq -80(%rsp), %rax # 8-byte Reload adcq -104(%rsp), %rax # 8-byte Folded Reload adcq $0, %r9 addq %rbx, %r11 adcq %rbp, %r10 adcq %r12, %r8 adcq -88(%rsp), %rdi # 8-byte Folded Reload adcq -72(%rsp), %rsi # 8-byte Folded Reload adcq -56(%rsp), %r13 # 8-byte Folded Reload adcq -64(%rsp), %rax # 8-byte Folded Reload movq %rax, -80(%rsp) # 8-byte Spill adcq $0, %r9 movq %r11, %rbx imulq 24(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -56(%rsp) # 8-byte Spill movq %rbx, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, -104(%rsp) # 8-byte Spill movq %rax, -64(%rsp) # 8-byte Spill movq %rbx, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -112(%rsp) # 8-byte Spill movq %rax, %r12 movq %rbx, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, %r15 movq %rbx, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, -120(%rsp) # 8-byte Spill movq %rax, %rbp movq %rbx, %rax mulq -8(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %rcx movq %rbx, %rax mulq 32(%rsp) # 8-byte Folded Reload addq %r11, %rax adcq %r10, %rcx adcq %r8, %rbp adcq %rdi, %r15 adcq %rsi, %r12 movq -64(%rsp), %rsi # 8-byte Reload adcq %r13, %rsi movq -56(%rsp), %rax # 8-byte Reload adcq -80(%rsp), %rax # 8-byte Folded Reload adcq $0, %r9 addq %rdx, %rcx adcq %r14, %rbp adcq -120(%rsp), %r15 # 8-byte Folded Reload adcq -72(%rsp), %r12 # 8-byte Folded Reload movq %r12, -72(%rsp) # 8-byte Spill adcq -112(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -64(%rsp) # 8-byte Spill adcq -104(%rsp), %rax # 8-byte Folded Reload movq %rax, -56(%rsp) # 8-byte Spill adcq -88(%rsp), %r9 # 8-byte Folded Reload movq 16(%rsp), %rax # 8-byte Reload movq 48(%rax), %rdi movq %rdi, %rax mulq 8(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, -80(%rsp) # 8-byte Spill movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, 16(%rsp) # 8-byte Spill movq %rax, (%rsp) # 8-byte Spill movq %rdi, %rax mulq -48(%rsp) # 8-byte Folded Reload movq %rdx, 8(%rsp) # 8-byte Spill movq %rax, -48(%rsp) # 8-byte Spill movq %rdi, %rax mulq -40(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, -40(%rsp) # 8-byte Spill movq %rdi, %rax mulq -32(%rsp) # 8-byte Folded Reload movq %rdx, %r11 movq %rax, %rsi movq %rdi, %rax mulq -24(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, %rbx movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rax, %r12 movq %rdx, %r8 addq %rbx, %r8 adcq %rsi, %r10 adcq -40(%rsp), %r11 # 8-byte Folded Reload adcq -48(%rsp), %r13 # 8-byte Folded Reload movq 8(%rsp), %rdx # 8-byte Reload adcq (%rsp), %rdx # 8-byte Folded Reload movq 16(%rsp), %rax # 8-byte Reload adcq -80(%rsp), %rax # 8-byte Folded Reload adcq $0, %r14 addq %rcx, %r12 adcq %rbp, %r8 adcq %r15, %r10 adcq -72(%rsp), %r11 # 8-byte Folded Reload adcq -64(%rsp), %r13 # 8-byte Folded Reload adcq -56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 8(%rsp) # 8-byte Spill adcq %r9, %rax movq %rax, 16(%rsp) # 8-byte Spill adcq $0, %r14 movq 24(%rsp), %rdi # 8-byte Reload imulq %r12, %rdi movq %rdi, %rax mulq 72(%rsp) # 8-byte Folded Reload movq %rdx, 24(%rsp) # 8-byte Spill movq %rax, %r9 movq %rdi, %rax mulq 64(%rsp) # 8-byte Folded Reload movq %rdx, (%rsp) # 8-byte Spill movq %rax, %rbp movq %rdi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -16(%rsp) # 8-byte Spill movq %rax, %rsi movq %rdi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, %rcx movq %rdi, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, %r15 movq %rdi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, %rbx movq %rdi, %rax movq -8(%rsp), %rdi # 8-byte Reload mulq %rdi addq %r12, %r15 adcq %r8, %rax adcq %r10, %rbx adcq %r11, %rcx adcq %r13, %rsi adcq 8(%rsp), %rbp # 8-byte Folded Reload adcq 16(%rsp), %r9 # 8-byte Folded Reload adcq $0, %r14 addq -32(%rsp), %rax # 8-byte Folded Reload adcq %rdx, %rbx adcq -40(%rsp), %rcx # 8-byte Folded Reload adcq -24(%rsp), %rsi # 8-byte Folded Reload adcq -16(%rsp), %rbp # 8-byte Folded Reload adcq (%rsp), %r9 # 8-byte Folded Reload adcq 24(%rsp), %r14 # 8-byte Folded Reload movq %rax, %r13 subq 32(%rsp), %r13 # 8-byte Folded Reload movq %rbx, %r12 sbbq %rdi, %r12 movq %rcx, %r8 sbbq 40(%rsp), %r8 # 8-byte Folded Reload movq %rsi, %r10 sbbq 48(%rsp), %r10 # 8-byte Folded Reload movq %rbp, %r11 sbbq 56(%rsp), %r11 # 8-byte Folded Reload movq %r9, %r15 sbbq 64(%rsp), %r15 # 8-byte Folded Reload movq %r14, %rdx sbbq 72(%rsp), %rdx # 8-byte Folded Reload movq %rdx, %rdi sarq $63, %rdi cmovsq %rax, %r13 movq -96(%rsp), %rax # 8-byte Reload movq %r13, (%rax) cmovsq %rbx, %r12 movq %r12, 8(%rax) cmovsq %rcx, %r8 movq %r8, 16(%rax) cmovsq %rsi, %r10 movq %r10, 24(%rax) cmovsq %rbp, %r11 movq %r11, 32(%rax) cmovsq %r9, %r15 movq %r15, 40(%rax) cmovsq %r14, %rdx movq %rdx, 48(%rax) addq $80, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end99: .size mcl_fp_montNF7L, .Lfunc_end99-mcl_fp_montNF7L .globl mcl_fp_montRed7L .align 16, 0x90 .type mcl_fp_montRed7L,@function mcl_fp_montRed7L: # @mcl_fp_montRed7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $64, %rsp movq %rdx, %rcx movq %rdi, -104(%rsp) # 8-byte Spill movq -8(%rcx), %rax movq %rax, 8(%rsp) # 8-byte Spill movq (%rcx), %rdx movq %rdx, 32(%rsp) # 8-byte Spill movq (%rsi), %rbp movq %rbp, 24(%rsp) # 8-byte Spill imulq %rax, %rbp movq 48(%rcx), %rdx movq %rdx, -16(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rax, -8(%rsp) # 8-byte Spill movq %rdx, -48(%rsp) # 8-byte Spill movq 40(%rcx), %rdx movq %rdx, (%rsp) # 8-byte Spill movq 32(%rcx), %r10 movq %r10, 56(%rsp) # 8-byte Spill movq 24(%rcx), %rdi movq %rdi, 48(%rsp) # 8-byte Spill movq 16(%rcx), %rbx movq %rbx, 40(%rsp) # 8-byte Spill movq 8(%rcx), %rcx movq %rcx, 16(%rsp) # 8-byte Spill movq %rbp, %rax mulq %rdx movq %rdx, %r13 movq %rax, %r9 movq %rbp, %rax mulq %r10 movq %rdx, %r15 movq %rax, %r11 movq %rbp, %rax mulq %rdi movq %rdx, %r10 movq %rax, %r8 movq %rbp, %rax mulq %rbx movq %rdx, %r14 movq %rax, %rbx movq %rbp, %rax mulq %rcx movq %rdx, %r12 movq %rax, %rdi movq %rbp, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %rbp addq %rdi, %rbp adcq %rbx, %r12 adcq %r8, %r14 adcq %r11, %r10 adcq %r9, %r15 adcq -8(%rsp), %r13 # 8-byte Folded Reload movq -48(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq 24(%rsp), %rax # 8-byte Folded Reload adcq 8(%rsi), %rbp adcq 16(%rsi), %r12 adcq 24(%rsi), %r14 adcq 32(%rsi), %r10 adcq 40(%rsi), %r15 adcq 48(%rsi), %r13 movq %r13, -80(%rsp) # 8-byte Spill adcq 56(%rsi), %rdx movq %rdx, -48(%rsp) # 8-byte Spill movq 104(%rsi), %r8 movq 96(%rsi), %rdx movq 88(%rsi), %rdi movq 80(%rsi), %rbx movq 72(%rsi), %rax movq 64(%rsi), %rsi adcq $0, %rsi movq %rsi, -88(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, -96(%rsp) # 8-byte Spill adcq $0, %rbx movq %rbx, -40(%rsp) # 8-byte Spill adcq $0, %rdi movq %rdi, -32(%rsp) # 8-byte Spill adcq $0, %rdx movq %rdx, -24(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, -8(%rsp) # 8-byte Spill sbbq %rax, %rax andl $1, %eax movq %rax, 24(%rsp) # 8-byte Spill movq %rbp, %rdi imulq 8(%rsp), %rdi # 8-byte Folded Reload movq %rdi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rdi, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rdi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -128(%rsp) # 8-byte Spill movq %rdi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %rbx movq %rdi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r11 movq %rdi, %rax mulq %rcx movq %rdx, %r8 movq %rax, %rcx movq %rdi, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r9 addq %rcx, %r9 adcq %r11, %r8 adcq %rbx, %rsi adcq -128(%rsp), %r13 # 8-byte Folded Reload movq -72(%rsp), %rdi # 8-byte Reload adcq -120(%rsp), %rdi # 8-byte Folded Reload movq -64(%rsp), %rdx # 8-byte Reload adcq -112(%rsp), %rdx # 8-byte Folded Reload movq -56(%rsp), %rcx # 8-byte Reload adcq $0, %rcx addq %rbp, %rax adcq %r12, %r9 adcq %r14, %r8 adcq %r10, %rsi adcq %r15, %r13 adcq -80(%rsp), %rdi # 8-byte Folded Reload movq %rdi, -72(%rsp) # 8-byte Spill adcq -48(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill adcq -88(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -56(%rsp) # 8-byte Spill adcq $0, -96(%rsp) # 8-byte Folded Spill adcq $0, -40(%rsp) # 8-byte Folded Spill adcq $0, -32(%rsp) # 8-byte Folded Spill adcq $0, -24(%rsp) # 8-byte Folded Spill movq -8(%rsp), %rbx # 8-byte Reload adcq $0, %rbx adcq $0, 24(%rsp) # 8-byte Folded Spill movq %r9, %rcx imulq 8(%rsp), %rcx # 8-byte Folded Reload movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -8(%rsp) # 8-byte Spill movq %rcx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rcx, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -88(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rcx, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, %r11 movq %rcx, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %rbp movq %rax, %r12 movq %rcx, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r14 movq %rcx, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r10 addq %r14, %r10 adcq %r12, %rdi adcq %r11, %rbp adcq -120(%rsp), %r15 # 8-byte Folded Reload movq -88(%rsp), %r11 # 8-byte Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload movq -80(%rsp), %rdx # 8-byte Reload adcq -8(%rsp), %rdx # 8-byte Folded Reload movq -48(%rsp), %rcx # 8-byte Reload adcq $0, %rcx addq %r9, %rax adcq %r8, %r10 adcq %rsi, %rdi adcq %r13, %rbp adcq -72(%rsp), %r15 # 8-byte Folded Reload adcq -64(%rsp), %r11 # 8-byte Folded Reload movq %r11, -88(%rsp) # 8-byte Spill adcq -56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -80(%rsp) # 8-byte Spill adcq -96(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -48(%rsp) # 8-byte Spill adcq $0, -40(%rsp) # 8-byte Folded Spill adcq $0, -32(%rsp) # 8-byte Folded Spill adcq $0, -24(%rsp) # 8-byte Folded Spill adcq $0, %rbx movq %rbx, -8(%rsp) # 8-byte Spill adcq $0, 24(%rsp) # 8-byte Folded Spill movq %r10, %rbx imulq 8(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %rax movq -16(%rsp), %r12 # 8-byte Reload mulq %r12 movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -96(%rsp) # 8-byte Spill movq %rbx, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -112(%rsp) # 8-byte Spill movq %rbx, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -72(%rsp) # 8-byte Spill movq %rax, -120(%rsp) # 8-byte Spill movq %rbx, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r8 movq %rax, %r14 movq %rbx, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, %r13 movq %rbx, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rcx movq %rax, %r11 movq %rbx, %rax mulq 32(%rsp) # 8-byte Folded Reload movq %rdx, %r9 addq %r11, %r9 adcq %r13, %rcx adcq %r14, %rsi adcq -120(%rsp), %r8 # 8-byte Folded Reload movq -72(%rsp), %r11 # 8-byte Reload adcq -112(%rsp), %r11 # 8-byte Folded Reload movq -64(%rsp), %rbx # 8-byte Reload adcq -96(%rsp), %rbx # 8-byte Folded Reload movq -56(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r10, %rax adcq %rdi, %r9 adcq %rbp, %rcx adcq %r15, %rsi adcq -88(%rsp), %r8 # 8-byte Folded Reload adcq -80(%rsp), %r11 # 8-byte Folded Reload movq %r11, -72(%rsp) # 8-byte Spill adcq -48(%rsp), %rbx # 8-byte Folded Reload movq %rbx, -64(%rsp) # 8-byte Spill adcq -40(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill adcq $0, -32(%rsp) # 8-byte Folded Spill adcq $0, -24(%rsp) # 8-byte Folded Spill adcq $0, -8(%rsp) # 8-byte Folded Spill adcq $0, 24(%rsp) # 8-byte Folded Spill movq %r9, %rbp imulq 8(%rsp), %rbp # 8-byte Folded Reload movq %rbp, %rax mulq %r12 movq %rdx, -40(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rbp, %rax mulq (%rsp) # 8-byte Folded Reload movq %rdx, -48(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rbp, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -96(%rsp) # 8-byte Spill movq %rbp, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r13 movq %rax, %r14 movq %rbp, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r11 movq %rbp, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r12 movq %rbp, %rax movq 32(%rsp), %rbp # 8-byte Reload mulq %rbp movq %rdx, %r10 addq %r12, %r10 adcq %r11, %rbx adcq %r14, %rdi adcq -96(%rsp), %r13 # 8-byte Folded Reload adcq -88(%rsp), %r15 # 8-byte Folded Reload movq -48(%rsp), %r11 # 8-byte Reload adcq -80(%rsp), %r11 # 8-byte Folded Reload movq -40(%rsp), %rdx # 8-byte Reload adcq $0, %rdx addq %r9, %rax adcq %rcx, %r10 adcq %rsi, %rbx adcq %r8, %rdi adcq -72(%rsp), %r13 # 8-byte Folded Reload adcq -64(%rsp), %r15 # 8-byte Folded Reload adcq -56(%rsp), %r11 # 8-byte Folded Reload movq %r11, -48(%rsp) # 8-byte Spill adcq -32(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -40(%rsp) # 8-byte Spill adcq $0, -24(%rsp) # 8-byte Folded Spill adcq $0, -8(%rsp) # 8-byte Folded Spill adcq $0, 24(%rsp) # 8-byte Folded Spill movq %r10, %rsi imulq 8(%rsp), %rsi # 8-byte Folded Reload movq %rsi, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, -72(%rsp) # 8-byte Spill movq %rsi, %rax movq (%rsp), %r8 # 8-byte Reload mulq %r8 movq %rdx, -56(%rsp) # 8-byte Spill movq %rax, -80(%rsp) # 8-byte Spill movq %rsi, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, -64(%rsp) # 8-byte Spill movq %rax, -88(%rsp) # 8-byte Spill movq %rsi, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r12 movq %rax, -96(%rsp) # 8-byte Spill movq %rsi, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %r9 movq %rax, %r11 movq %rsi, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %r14 movq %rax, %rcx movq %rsi, %rax mulq %rbp movq %rdx, %rbp addq %rcx, %rbp adcq %r11, %r14 adcq -96(%rsp), %r9 # 8-byte Folded Reload adcq -88(%rsp), %r12 # 8-byte Folded Reload movq -64(%rsp), %rsi # 8-byte Reload adcq -80(%rsp), %rsi # 8-byte Folded Reload movq -56(%rsp), %rdx # 8-byte Reload adcq -72(%rsp), %rdx # 8-byte Folded Reload movq -32(%rsp), %rcx # 8-byte Reload adcq $0, %rcx addq %r10, %rax adcq %rbx, %rbp adcq %rdi, %r14 adcq %r13, %r9 adcq %r15, %r12 adcq -48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, -64(%rsp) # 8-byte Spill adcq -40(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -56(%rsp) # 8-byte Spill adcq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -32(%rsp) # 8-byte Spill adcq $0, -8(%rsp) # 8-byte Folded Spill adcq $0, 24(%rsp) # 8-byte Folded Spill movq 8(%rsp), %rcx # 8-byte Reload imulq %rbp, %rcx movq %rcx, %rax mulq -16(%rsp) # 8-byte Folded Reload movq %rdx, %rsi movq %rax, 8(%rsp) # 8-byte Spill movq %rcx, %rax mulq %r8 movq %rdx, %r13 movq %rax, -24(%rsp) # 8-byte Spill movq %rcx, %rax mulq 56(%rsp) # 8-byte Folded Reload movq %rdx, %r15 movq %rax, -40(%rsp) # 8-byte Spill movq %rcx, %rax mulq 48(%rsp) # 8-byte Folded Reload movq %rdx, %r10 movq %rax, -48(%rsp) # 8-byte Spill movq %rcx, %rax mulq 40(%rsp) # 8-byte Folded Reload movq %rdx, %rdi movq %rax, %r8 movq %rcx, %rax mulq 16(%rsp) # 8-byte Folded Reload movq %rdx, %rbx movq %rax, %r11 movq %rcx, %rax mulq 32(%rsp) # 8-byte Folded Reload addq %r11, %rdx adcq %r8, %rbx adcq -48(%rsp), %rdi # 8-byte Folded Reload adcq -40(%rsp), %r10 # 8-byte Folded Reload adcq -24(%rsp), %r15 # 8-byte Folded Reload adcq 8(%rsp), %r13 # 8-byte Folded Reload adcq $0, %rsi addq %rbp, %rax adcq %r14, %rdx adcq %r9, %rbx adcq %r12, %rdi adcq -64(%rsp), %r10 # 8-byte Folded Reload adcq -56(%rsp), %r15 # 8-byte Folded Reload adcq -32(%rsp), %r13 # 8-byte Folded Reload adcq -8(%rsp), %rsi # 8-byte Folded Reload movq 24(%rsp), %rcx # 8-byte Reload adcq $0, %rcx movq %rdx, %rax subq 32(%rsp), %rax # 8-byte Folded Reload movq %rbx, %rbp sbbq 16(%rsp), %rbp # 8-byte Folded Reload movq %rdi, %r8 sbbq 40(%rsp), %r8 # 8-byte Folded Reload movq %r10, %r9 sbbq 48(%rsp), %r9 # 8-byte Folded Reload movq %r15, %r11 sbbq 56(%rsp), %r11 # 8-byte Folded Reload movq %r13, %r14 sbbq (%rsp), %r14 # 8-byte Folded Reload movq %rsi, %r12 sbbq -16(%rsp), %r12 # 8-byte Folded Reload sbbq $0, %rcx andl $1, %ecx cmovneq %rsi, %r12 testb %cl, %cl cmovneq %rdx, %rax movq -104(%rsp), %rcx # 8-byte Reload movq %rax, (%rcx) cmovneq %rbx, %rbp movq %rbp, 8(%rcx) cmovneq %rdi, %r8 movq %r8, 16(%rcx) cmovneq %r10, %r9 movq %r9, 24(%rcx) cmovneq %r15, %r11 movq %r11, 32(%rcx) cmovneq %r13, %r14 movq %r14, 40(%rcx) movq %r12, 48(%rcx) addq $64, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end100: .size mcl_fp_montRed7L, .Lfunc_end100-mcl_fp_montRed7L .globl mcl_fp_addPre7L .align 16, 0x90 .type mcl_fp_addPre7L,@function mcl_fp_addPre7L: # @mcl_fp_addPre7L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq 48(%rdx), %r8 movq 48(%rsi), %r14 movq 40(%rdx), %r9 movq 40(%rsi), %r15 movq 32(%rdx), %r10 movq 24(%rdx), %r11 movq 16(%rdx), %r12 movq (%rdx), %rcx movq 8(%rdx), %rdx addq (%rsi), %rcx adcq 8(%rsi), %rdx movq 24(%rsi), %rax movq 32(%rsi), %rbx adcq 16(%rsi), %r12 movq %rcx, (%rdi) movq %rdx, 8(%rdi) movq %r12, 16(%rdi) adcq %r11, %rax movq %rax, 24(%rdi) adcq %r10, %rbx movq %rbx, 32(%rdi) adcq %r9, %r15 movq %r15, 40(%rdi) adcq %r8, %r14 movq %r14, 48(%rdi) sbbq %rax, %rax andl $1, %eax popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end101: .size mcl_fp_addPre7L, .Lfunc_end101-mcl_fp_addPre7L .globl mcl_fp_subPre7L .align 16, 0x90 .type mcl_fp_subPre7L,@function mcl_fp_subPre7L: # @mcl_fp_subPre7L # BB#0: pushq %r15 pushq %r14 pushq %r12 pushq %rbx movq 48(%rdx), %r8 movq 48(%rsi), %r10 movq 40(%rdx), %r9 movq 40(%rsi), %r15 movq 24(%rdx), %r11 movq 32(%rdx), %r14 movq (%rsi), %rbx movq 8(%rsi), %r12 xorl %eax, %eax subq (%rdx), %rbx sbbq 8(%rdx), %r12 movq 16(%rsi), %rcx sbbq 16(%rdx), %rcx movq 32(%rsi), %rdx movq 24(%rsi), %rsi movq %rbx, (%rdi) movq %r12, 8(%rdi) movq %rcx, 16(%rdi) sbbq %r11, %rsi movq %rsi, 24(%rdi) sbbq %r14, %rdx movq %rdx, 32(%rdi) sbbq %r9, %r15 movq %r15, 40(%rdi) sbbq %r8, %r10 movq %r10, 48(%rdi) sbbq $0, %rax andl $1, %eax popq %rbx popq %r12 popq %r14 popq %r15 retq .Lfunc_end102: .size mcl_fp_subPre7L, .Lfunc_end102-mcl_fp_subPre7L .globl mcl_fp_shr1_7L .align 16, 0x90 .type mcl_fp_shr1_7L,@function mcl_fp_shr1_7L: # @mcl_fp_shr1_7L # BB#0: movq 48(%rsi), %r8 movq 40(%rsi), %r9 movq 32(%rsi), %r10 movq 24(%rsi), %rax movq 16(%rsi), %rcx movq (%rsi), %rdx movq 8(%rsi), %rsi shrdq $1, %rsi, %rdx movq %rdx, (%rdi) shrdq $1, %rcx, %rsi movq %rsi, 8(%rdi) shrdq $1, %rax, %rcx movq %rcx, 16(%rdi) shrdq $1, %r10, %rax movq %rax, 24(%rdi) shrdq $1, %r9, %r10 movq %r10, 32(%rdi) shrdq $1, %r8, %r9 movq %r9, 40(%rdi) shrq %r8 movq %r8, 48(%rdi) retq .Lfunc_end103: .size mcl_fp_shr1_7L, .Lfunc_end103-mcl_fp_shr1_7L .globl mcl_fp_add7L .align 16, 0x90 .type mcl_fp_add7L,@function mcl_fp_add7L: # @mcl_fp_add7L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 48(%rdx), %r14 movq 48(%rsi), %r8 movq 40(%rdx), %r15 movq 40(%rsi), %r9 movq 32(%rdx), %r12 movq 24(%rdx), %r13 movq 16(%rdx), %r10 movq (%rdx), %r11 movq 8(%rdx), %rdx addq (%rsi), %r11 adcq 8(%rsi), %rdx movq 24(%rsi), %rax movq 32(%rsi), %rbx adcq 16(%rsi), %r10 movq %r11, (%rdi) movq %rdx, 8(%rdi) movq %r10, 16(%rdi) adcq %r13, %rax movq %rax, 24(%rdi) adcq %r12, %rbx movq %rbx, 32(%rdi) adcq %r15, %r9 movq %r9, 40(%rdi) adcq %r14, %r8 movq %r8, 48(%rdi) sbbq %rsi, %rsi andl $1, %esi subq (%rcx), %r11 sbbq 8(%rcx), %rdx sbbq 16(%rcx), %r10 sbbq 24(%rcx), %rax sbbq 32(%rcx), %rbx sbbq 40(%rcx), %r9 sbbq 48(%rcx), %r8 sbbq $0, %rsi testb $1, %sil jne .LBB104_2 # BB#1: # %nocarry movq %r11, (%rdi) movq %rdx, 8(%rdi) movq %r10, 16(%rdi) movq %rax, 24(%rdi) movq %rbx, 32(%rdi) movq %r9, 40(%rdi) movq %r8, 48(%rdi) .LBB104_2: # %carry popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end104: .size mcl_fp_add7L, .Lfunc_end104-mcl_fp_add7L .globl mcl_fp_addNF7L .align 16, 0x90 .type mcl_fp_addNF7L,@function mcl_fp_addNF7L: # @mcl_fp_addNF7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 48(%rdx), %r9 movq 40(%rdx), %rbp movq 32(%rdx), %r10 movq 24(%rdx), %r11 movq 16(%rdx), %r14 movq (%rdx), %r12 movq 8(%rdx), %r15 addq (%rsi), %r12 adcq 8(%rsi), %r15 adcq 16(%rsi), %r14 adcq 24(%rsi), %r11 adcq 32(%rsi), %r10 adcq 40(%rsi), %rbp movq %rbp, -8(%rsp) # 8-byte Spill adcq 48(%rsi), %r9 movq %r12, %rsi subq (%rcx), %rsi movq %r15, %rdx sbbq 8(%rcx), %rdx movq %r14, %rax sbbq 16(%rcx), %rax movq %r11, %rbx sbbq 24(%rcx), %rbx movq %r10, %r13 sbbq 32(%rcx), %r13 sbbq 40(%rcx), %rbp movq %r9, %r8 sbbq 48(%rcx), %r8 movq %r8, %rcx sarq $63, %rcx cmovsq %r12, %rsi movq %rsi, (%rdi) cmovsq %r15, %rdx movq %rdx, 8(%rdi) cmovsq %r14, %rax movq %rax, 16(%rdi) cmovsq %r11, %rbx movq %rbx, 24(%rdi) cmovsq %r10, %r13 movq %r13, 32(%rdi) cmovsq -8(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 40(%rdi) cmovsq %r9, %r8 movq %r8, 48(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end105: .size mcl_fp_addNF7L, .Lfunc_end105-mcl_fp_addNF7L .globl mcl_fp_sub7L .align 16, 0x90 .type mcl_fp_sub7L,@function mcl_fp_sub7L: # @mcl_fp_sub7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 48(%rdx), %r14 movq 48(%rsi), %r8 movq 40(%rdx), %r15 movq 40(%rsi), %r9 movq 32(%rdx), %r12 movq (%rsi), %rax movq 8(%rsi), %r11 xorl %ebx, %ebx subq (%rdx), %rax sbbq 8(%rdx), %r11 movq 16(%rsi), %r13 sbbq 16(%rdx), %r13 movq 32(%rsi), %r10 movq 24(%rsi), %rsi sbbq 24(%rdx), %rsi movq %rax, (%rdi) movq %r11, 8(%rdi) movq %r13, 16(%rdi) movq %rsi, 24(%rdi) sbbq %r12, %r10 movq %r10, 32(%rdi) sbbq %r15, %r9 movq %r9, 40(%rdi) sbbq %r14, %r8 movq %r8, 48(%rdi) sbbq $0, %rbx testb $1, %bl je .LBB106_2 # BB#1: # %carry movq 48(%rcx), %r14 movq 40(%rcx), %r15 movq 32(%rcx), %r12 movq 24(%rcx), %rbx movq 8(%rcx), %rdx movq 16(%rcx), %rbp addq (%rcx), %rax movq %rax, (%rdi) adcq %r11, %rdx movq %rdx, 8(%rdi) adcq %r13, %rbp movq %rbp, 16(%rdi) adcq %rsi, %rbx movq %rbx, 24(%rdi) adcq %r10, %r12 movq %r12, 32(%rdi) adcq %r9, %r15 movq %r15, 40(%rdi) adcq %r8, %r14 movq %r14, 48(%rdi) .LBB106_2: # %nocarry popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end106: .size mcl_fp_sub7L, .Lfunc_end106-mcl_fp_sub7L .globl mcl_fp_subNF7L .align 16, 0x90 .type mcl_fp_subNF7L,@function mcl_fp_subNF7L: # @mcl_fp_subNF7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r8 movq 48(%rsi), %r12 movq 40(%rsi), %rax movq 32(%rsi), %r9 movq 24(%rsi), %r10 movq 16(%rsi), %r11 movq (%rsi), %r14 movq 8(%rsi), %r15 subq (%rdx), %r14 sbbq 8(%rdx), %r15 sbbq 16(%rdx), %r11 sbbq 24(%rdx), %r10 sbbq 32(%rdx), %r9 sbbq 40(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill sbbq 48(%rdx), %r12 movq %r12, %rax sarq $63, %rax movq %rax, %rsi shldq $1, %r12, %rsi andq (%r8), %rsi movq 48(%r8), %r13 andq %rax, %r13 movq 40(%r8), %rbx andq %rax, %rbx movq 32(%r8), %rdx andq %rax, %rdx movq 24(%r8), %rbp andq %rax, %rbp movq 16(%r8), %rcx andq %rax, %rcx andq 8(%r8), %rax addq %r14, %rsi adcq %r15, %rax movq %rsi, (%rdi) movq %rax, 8(%rdi) adcq %r11, %rcx movq %rcx, 16(%rdi) adcq %r10, %rbp movq %rbp, 24(%rdi) adcq %r9, %rdx movq %rdx, 32(%rdi) adcq -8(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 40(%rdi) adcq %r12, %r13 movq %r13, 48(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end107: .size mcl_fp_subNF7L, .Lfunc_end107-mcl_fp_subNF7L .globl mcl_fpDbl_add7L .align 16, 0x90 .type mcl_fpDbl_add7L,@function mcl_fpDbl_add7L: # @mcl_fpDbl_add7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r8 movq 104(%rdx), %rax movq %rax, -16(%rsp) # 8-byte Spill movq 96(%rdx), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 88(%rdx), %r11 movq 80(%rdx), %r14 movq 24(%rsi), %r15 movq 32(%rsi), %r12 movq 16(%rdx), %r9 movq (%rdx), %rax movq 8(%rdx), %rbx addq (%rsi), %rax adcq 8(%rsi), %rbx adcq 16(%rsi), %r9 adcq 24(%rdx), %r15 adcq 32(%rdx), %r12 movq 72(%rdx), %r13 movq 64(%rdx), %rbp movq %rax, (%rdi) movq 56(%rdx), %r10 movq %rbx, 8(%rdi) movq 48(%rdx), %rcx movq 40(%rdx), %rdx movq %r9, 16(%rdi) movq 104(%rsi), %r9 movq %r15, 24(%rdi) movq 40(%rsi), %rbx adcq %rdx, %rbx movq 96(%rsi), %r15 movq %r12, 32(%rdi) movq 48(%rsi), %rdx adcq %rcx, %rdx movq 88(%rsi), %rax movq %rbx, 40(%rdi) movq 56(%rsi), %rcx adcq %r10, %rcx movq 80(%rsi), %r12 movq %rdx, 48(%rdi) movq 72(%rsi), %rdx movq 64(%rsi), %rsi adcq %rbp, %rsi adcq %r13, %rdx adcq %r14, %r12 adcq %r11, %rax movq %rax, -8(%rsp) # 8-byte Spill adcq -24(%rsp), %r15 # 8-byte Folded Reload movq %r15, -24(%rsp) # 8-byte Spill adcq -16(%rsp), %r9 # 8-byte Folded Reload sbbq %rbp, %rbp andl $1, %ebp movq %rcx, %rbx subq (%r8), %rbx movq %rsi, %r10 sbbq 8(%r8), %r10 movq %rdx, %r11 sbbq 16(%r8), %r11 movq %r12, %r14 sbbq 24(%r8), %r14 movq -8(%rsp), %r13 # 8-byte Reload sbbq 32(%r8), %r13 sbbq 40(%r8), %r15 movq %r9, %rax sbbq 48(%r8), %rax sbbq $0, %rbp andl $1, %ebp cmovneq %rcx, %rbx movq %rbx, 56(%rdi) testb %bpl, %bpl cmovneq %rsi, %r10 movq %r10, 64(%rdi) cmovneq %rdx, %r11 movq %r11, 72(%rdi) cmovneq %r12, %r14 movq %r14, 80(%rdi) cmovneq -8(%rsp), %r13 # 8-byte Folded Reload movq %r13, 88(%rdi) cmovneq -24(%rsp), %r15 # 8-byte Folded Reload movq %r15, 96(%rdi) cmovneq %r9, %rax movq %rax, 104(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end108: .size mcl_fpDbl_add7L, .Lfunc_end108-mcl_fpDbl_add7L .globl mcl_fpDbl_sub7L .align 16, 0x90 .type mcl_fpDbl_sub7L,@function mcl_fpDbl_sub7L: # @mcl_fpDbl_sub7L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r8 movq 104(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 96(%rdx), %r10 movq 88(%rdx), %r14 movq 16(%rsi), %rax movq (%rsi), %r15 movq 8(%rsi), %r11 xorl %ecx, %ecx subq (%rdx), %r15 sbbq 8(%rdx), %r11 sbbq 16(%rdx), %rax movq 24(%rsi), %rbx sbbq 24(%rdx), %rbx movq 32(%rsi), %r12 sbbq 32(%rdx), %r12 movq 80(%rdx), %r13 movq 72(%rdx), %rbp movq %r15, (%rdi) movq 64(%rdx), %r9 movq %r11, 8(%rdi) movq 56(%rdx), %r15 movq %rax, 16(%rdi) movq 48(%rdx), %r11 movq 40(%rdx), %rdx movq %rbx, 24(%rdi) movq 40(%rsi), %rbx sbbq %rdx, %rbx movq 104(%rsi), %rax movq %r12, 32(%rdi) movq 48(%rsi), %r12 sbbq %r11, %r12 movq 96(%rsi), %r11 movq %rbx, 40(%rdi) movq 56(%rsi), %rdx sbbq %r15, %rdx movq 88(%rsi), %r15 movq %r12, 48(%rdi) movq 64(%rsi), %rbx sbbq %r9, %rbx movq 80(%rsi), %r12 movq 72(%rsi), %r9 sbbq %rbp, %r9 sbbq %r13, %r12 sbbq %r14, %r15 sbbq %r10, %r11 sbbq -8(%rsp), %rax # 8-byte Folded Reload movq %rax, -8(%rsp) # 8-byte Spill movl $0, %ebp sbbq $0, %rbp andl $1, %ebp movq (%r8), %r10 cmoveq %rcx, %r10 testb %bpl, %bpl movq 16(%r8), %rbp cmoveq %rcx, %rbp movq 8(%r8), %rsi cmoveq %rcx, %rsi movq 48(%r8), %r14 cmoveq %rcx, %r14 movq 40(%r8), %r13 cmoveq %rcx, %r13 movq 32(%r8), %rax cmoveq %rcx, %rax cmovneq 24(%r8), %rcx addq %rdx, %r10 adcq %rbx, %rsi movq %r10, 56(%rdi) movq %rsi, 64(%rdi) adcq %r9, %rbp movq %rbp, 72(%rdi) adcq %r12, %rcx movq %rcx, 80(%rdi) adcq %r15, %rax movq %rax, 88(%rdi) adcq %r11, %r13 movq %r13, 96(%rdi) adcq -8(%rsp), %r14 # 8-byte Folded Reload movq %r14, 104(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end109: .size mcl_fpDbl_sub7L, .Lfunc_end109-mcl_fpDbl_sub7L .align 16, 0x90 .type .LmulPv512x64,@function .LmulPv512x64: # @mulPv512x64 # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rcx movq %rcx, %rax mulq (%rsi) movq %rdx, -24(%rsp) # 8-byte Spill movq %rax, (%rdi) movq %rcx, %rax mulq 56(%rsi) movq %rdx, %r10 movq %rax, -8(%rsp) # 8-byte Spill movq %rcx, %rax mulq 48(%rsi) movq %rdx, %r11 movq %rax, -16(%rsp) # 8-byte Spill movq %rcx, %rax mulq 40(%rsi) movq %rdx, %r12 movq %rax, %r15 movq %rcx, %rax mulq 32(%rsi) movq %rdx, %rbx movq %rax, %r13 movq %rcx, %rax mulq 24(%rsi) movq %rdx, %rbp movq %rax, %r8 movq %rcx, %rax mulq 16(%rsi) movq %rdx, %r9 movq %rax, %r14 movq %rcx, %rax mulq 8(%rsi) addq -24(%rsp), %rax # 8-byte Folded Reload movq %rax, 8(%rdi) adcq %r14, %rdx movq %rdx, 16(%rdi) adcq %r8, %r9 movq %r9, 24(%rdi) adcq %r13, %rbp movq %rbp, 32(%rdi) adcq %r15, %rbx movq %rbx, 40(%rdi) adcq -16(%rsp), %r12 # 8-byte Folded Reload movq %r12, 48(%rdi) adcq -8(%rsp), %r11 # 8-byte Folded Reload movq %r11, 56(%rdi) adcq $0, %r10 movq %r10, 64(%rdi) movq %rdi, %rax popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end110: .size .LmulPv512x64, .Lfunc_end110-.LmulPv512x64 .globl mcl_fp_mulUnitPre8L .align 16, 0x90 .type mcl_fp_mulUnitPre8L,@function mcl_fp_mulUnitPre8L: # @mcl_fp_mulUnitPre8L # BB#0: pushq %rbx subq $80, %rsp movq %rdi, %rbx leaq 8(%rsp), %rdi callq .LmulPv512x64 movq 72(%rsp), %r8 movq 64(%rsp), %r9 movq 56(%rsp), %r10 movq 48(%rsp), %r11 movq 40(%rsp), %rdi movq 32(%rsp), %rax movq 24(%rsp), %rcx movq 8(%rsp), %rdx movq 16(%rsp), %rsi movq %rdx, (%rbx) movq %rsi, 8(%rbx) movq %rcx, 16(%rbx) movq %rax, 24(%rbx) movq %rdi, 32(%rbx) movq %r11, 40(%rbx) movq %r10, 48(%rbx) movq %r9, 56(%rbx) movq %r8, 64(%rbx) addq $80, %rsp popq %rbx retq .Lfunc_end111: .size mcl_fp_mulUnitPre8L, .Lfunc_end111-mcl_fp_mulUnitPre8L .globl mcl_fpDbl_mulPre8L .align 16, 0x90 .type mcl_fpDbl_mulPre8L,@function mcl_fpDbl_mulPre8L: # @mcl_fpDbl_mulPre8L # BB#0: pushq %rbp movq %rsp, %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $200, %rsp movq %rdx, %rbx movq %rsi, %r15 movq %rdi, %r14 callq mcl_fpDbl_mulPre4L@PLT leaq 64(%r14), %rdi leaq 32(%r15), %rsi leaq 32(%rbx), %rdx callq mcl_fpDbl_mulPre4L@PLT movq 56(%rbx), %r10 movq 48(%rbx), %rcx movq (%rbx), %rdx movq 8(%rbx), %rsi addq 32(%rbx), %rdx adcq 40(%rbx), %rsi adcq 16(%rbx), %rcx adcq 24(%rbx), %r10 pushfq popq %r8 xorl %r9d, %r9d movq 56(%r15), %rdi movq 48(%r15), %r13 movq (%r15), %r12 movq 8(%r15), %rbx addq 32(%r15), %r12 adcq 40(%r15), %rbx adcq 16(%r15), %r13 adcq 24(%r15), %rdi movl $0, %eax cmovbq %r10, %rax movq %rax, -176(%rbp) # 8-byte Spill movl $0, %eax cmovbq %rcx, %rax movq %rax, -184(%rbp) # 8-byte Spill movl $0, %eax cmovbq %rsi, %rax movq %rax, -192(%rbp) # 8-byte Spill movl $0, %eax cmovbq %rdx, %rax movq %rax, -200(%rbp) # 8-byte Spill sbbq %r15, %r15 movq %r12, -136(%rbp) movq %rbx, -128(%rbp) movq %r13, -120(%rbp) movq %rdi, -112(%rbp) movq %rdx, -168(%rbp) movq %rsi, -160(%rbp) movq %rcx, -152(%rbp) movq %r10, -144(%rbp) pushq %r8 popfq cmovaeq %r9, %rdi movq %rdi, -216(%rbp) # 8-byte Spill cmovaeq %r9, %r13 cmovaeq %r9, %rbx cmovaeq %r9, %r12 sbbq %rax, %rax movq %rax, -208(%rbp) # 8-byte Spill leaq -104(%rbp), %rdi leaq -136(%rbp), %rsi leaq -168(%rbp), %rdx callq mcl_fpDbl_mulPre4L@PLT addq -200(%rbp), %r12 # 8-byte Folded Reload adcq -192(%rbp), %rbx # 8-byte Folded Reload adcq -184(%rbp), %r13 # 8-byte Folded Reload movq -216(%rbp), %r10 # 8-byte Reload adcq -176(%rbp), %r10 # 8-byte Folded Reload sbbq %rax, %rax andl $1, %eax movq -208(%rbp), %rdx # 8-byte Reload andl %edx, %r15d andl $1, %r15d addq -72(%rbp), %r12 adcq -64(%rbp), %rbx adcq -56(%rbp), %r13 adcq -48(%rbp), %r10 adcq %rax, %r15 movq -80(%rbp), %rax movq -88(%rbp), %rcx movq -104(%rbp), %rsi movq -96(%rbp), %rdx subq (%r14), %rsi sbbq 8(%r14), %rdx sbbq 16(%r14), %rcx sbbq 24(%r14), %rax movq 32(%r14), %rdi movq %rdi, -184(%rbp) # 8-byte Spill movq 40(%r14), %r8 movq %r8, -176(%rbp) # 8-byte Spill sbbq %rdi, %r12 sbbq %r8, %rbx movq 48(%r14), %rdi movq %rdi, -192(%rbp) # 8-byte Spill sbbq %rdi, %r13 movq 56(%r14), %rdi movq %rdi, -200(%rbp) # 8-byte Spill sbbq %rdi, %r10 sbbq $0, %r15 movq 64(%r14), %r11 subq %r11, %rsi movq 72(%r14), %rdi movq %rdi, -208(%rbp) # 8-byte Spill sbbq %rdi, %rdx movq 80(%r14), %rdi movq %rdi, -216(%rbp) # 8-byte Spill sbbq %rdi, %rcx movq 88(%r14), %rdi movq %rdi, -224(%rbp) # 8-byte Spill sbbq %rdi, %rax movq 96(%r14), %rdi movq %rdi, -232(%rbp) # 8-byte Spill sbbq %rdi, %r12 movq 104(%r14), %rdi sbbq %rdi, %rbx movq 112(%r14), %r8 sbbq %r8, %r13 movq 120(%r14), %r9 sbbq %r9, %r10 sbbq $0, %r15 addq -184(%rbp), %rsi # 8-byte Folded Reload adcq -176(%rbp), %rdx # 8-byte Folded Reload movq %rsi, 32(%r14) adcq -192(%rbp), %rcx # 8-byte Folded Reload movq %rdx, 40(%r14) adcq -200(%rbp), %rax # 8-byte Folded Reload movq %rcx, 48(%r14) adcq %r11, %r12 movq %rax, 56(%r14) movq %r12, 64(%r14) adcq -208(%rbp), %rbx # 8-byte Folded Reload movq %rbx, 72(%r14) adcq -216(%rbp), %r13 # 8-byte Folded Reload movq %r13, 80(%r14) adcq -224(%rbp), %r10 # 8-byte Folded Reload movq %r10, 88(%r14) adcq -232(%rbp), %r15 # 8-byte Folded Reload movq %r15, 96(%r14) adcq $0, %rdi movq %rdi, 104(%r14) adcq $0, %r8 movq %r8, 112(%r14) adcq $0, %r9 movq %r9, 120(%r14) addq $200, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end112: .size mcl_fpDbl_mulPre8L, .Lfunc_end112-mcl_fpDbl_mulPre8L .globl mcl_fpDbl_sqrPre8L .align 16, 0x90 .type mcl_fpDbl_sqrPre8L,@function mcl_fpDbl_sqrPre8L: # @mcl_fpDbl_sqrPre8L # BB#0: pushq %rbp movq %rsp, %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $200, %rsp movq %rsi, %r14 movq %rdi, %rbx movq %r14, %rdx callq mcl_fpDbl_mulPre4L@PLT leaq 64(%rbx), %rdi leaq 32(%r14), %rsi movq %rsi, %rdx callq mcl_fpDbl_mulPre4L@PLT movq (%r14), %r12 movq 8(%r14), %r15 addq 32(%r14), %r12 adcq 40(%r14), %r15 pushfq popq %rax movq %r12, -136(%rbp) movq %r12, -168(%rbp) addq %r12, %r12 movq %r15, -128(%rbp) movq %r15, -160(%rbp) adcq %r15, %r15 pushfq popq %rcx movq 56(%r14), %r13 movq 48(%r14), %rdx pushq %rax popfq adcq 16(%r14), %rdx adcq 24(%r14), %r13 pushfq popq %r8 pushfq popq %rsi pushfq popq %rdi sbbq %rax, %rax movq %rax, -184(%rbp) # 8-byte Spill xorl %eax, %eax pushq %rdi popfq cmovaeq %rax, %r15 movq %r15, -176(%rbp) # 8-byte Spill cmovaeq %rax, %r12 movq %rdx, -120(%rbp) movq %rdx, -152(%rbp) movq %rdx, %r15 pushq %rcx popfq adcq %r15, %r15 movq %r13, %r14 movq %r13, -112(%rbp) movq %r13, -144(%rbp) adcq %r13, %r13 pushq %rsi popfq cmovaeq %rax, %r13 cmovaeq %rax, %r15 shrq $63, %r14 pushq %r8 popfq cmovaeq %rax, %r14 leaq -104(%rbp), %rdi leaq -136(%rbp), %rsi leaq -168(%rbp), %rdx callq mcl_fpDbl_mulPre4L@PLT movq -184(%rbp), %rax # 8-byte Reload andl $1, %eax addq -72(%rbp), %r12 movq -176(%rbp), %r8 # 8-byte Reload adcq -64(%rbp), %r8 adcq -56(%rbp), %r15 adcq -48(%rbp), %r13 adcq %r14, %rax movq %rax, %rdi movq -80(%rbp), %rax movq -88(%rbp), %rcx movq -104(%rbp), %rsi movq -96(%rbp), %rdx subq (%rbx), %rsi sbbq 8(%rbx), %rdx sbbq 16(%rbx), %rcx sbbq 24(%rbx), %rax movq 32(%rbx), %r10 movq %r10, -184(%rbp) # 8-byte Spill movq 40(%rbx), %r9 movq %r9, -176(%rbp) # 8-byte Spill sbbq %r10, %r12 sbbq %r9, %r8 movq %r8, %r10 movq 48(%rbx), %r8 movq %r8, -192(%rbp) # 8-byte Spill sbbq %r8, %r15 movq 56(%rbx), %r8 movq %r8, -200(%rbp) # 8-byte Spill sbbq %r8, %r13 sbbq $0, %rdi movq 64(%rbx), %r11 subq %r11, %rsi movq 72(%rbx), %r8 movq %r8, -208(%rbp) # 8-byte Spill sbbq %r8, %rdx movq 80(%rbx), %r8 movq %r8, -216(%rbp) # 8-byte Spill sbbq %r8, %rcx movq 88(%rbx), %r8 movq %r8, -224(%rbp) # 8-byte Spill sbbq %r8, %rax movq 96(%rbx), %r8 movq %r8, -232(%rbp) # 8-byte Spill sbbq %r8, %r12 movq 104(%rbx), %r14 sbbq %r14, %r10 movq 112(%rbx), %r8 sbbq %r8, %r15 movq 120(%rbx), %r9 sbbq %r9, %r13 sbbq $0, %rdi addq -184(%rbp), %rsi # 8-byte Folded Reload adcq -176(%rbp), %rdx # 8-byte Folded Reload movq %rsi, 32(%rbx) adcq -192(%rbp), %rcx # 8-byte Folded Reload movq %rdx, 40(%rbx) adcq -200(%rbp), %rax # 8-byte Folded Reload movq %rcx, 48(%rbx) adcq %r11, %r12 movq %rax, 56(%rbx) movq %r12, 64(%rbx) adcq -208(%rbp), %r10 # 8-byte Folded Reload movq %r10, 72(%rbx) adcq -216(%rbp), %r15 # 8-byte Folded Reload movq %r15, 80(%rbx) adcq -224(%rbp), %r13 # 8-byte Folded Reload movq %r13, 88(%rbx) adcq -232(%rbp), %rdi # 8-byte Folded Reload movq %rdi, 96(%rbx) adcq $0, %r14 movq %r14, 104(%rbx) adcq $0, %r8 movq %r8, 112(%rbx) adcq $0, %r9 movq %r9, 120(%rbx) addq $200, %rsp popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end113: .size mcl_fpDbl_sqrPre8L, .Lfunc_end113-mcl_fpDbl_sqrPre8L .globl mcl_fp_mont8L .align 16, 0x90 .type mcl_fp_mont8L,@function mcl_fp_mont8L: # @mcl_fp_mont8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $1256, %rsp # imm = 0x4E8 movq %rcx, %r13 movq %r13, 40(%rsp) # 8-byte Spill movq %rdx, 16(%rsp) # 8-byte Spill movq %rsi, 24(%rsp) # 8-byte Spill movq %rdi, (%rsp) # 8-byte Spill movq -8(%r13), %rbx movq %rbx, 32(%rsp) # 8-byte Spill movq (%rdx), %rdx leaq 1184(%rsp), %rdi callq .LmulPv512x64 movq 1184(%rsp), %r15 movq 1192(%rsp), %r14 movq %r15, %rdx imulq %rbx, %rdx movq 1248(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill movq 1240(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 1232(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 1224(%rsp), %rax movq %rax, 72(%rsp) # 8-byte Spill movq 1216(%rsp), %r12 movq 1208(%rsp), %rbx movq 1200(%rsp), %rbp leaq 1112(%rsp), %rdi movq %r13, %rsi callq .LmulPv512x64 addq 1112(%rsp), %r15 adcq 1120(%rsp), %r14 adcq 1128(%rsp), %rbp movq %rbp, 8(%rsp) # 8-byte Spill adcq 1136(%rsp), %rbx movq %rbx, 48(%rsp) # 8-byte Spill adcq 1144(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill movq 72(%rsp), %r13 # 8-byte Reload adcq 1152(%rsp), %r13 movq 88(%rsp), %rbx # 8-byte Reload adcq 1160(%rsp), %rbx movq 80(%rsp), %rbp # 8-byte Reload adcq 1168(%rsp), %rbp movq 96(%rsp), %rax # 8-byte Reload adcq 1176(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill sbbq %r15, %r15 movq 16(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdx leaq 1040(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 andl $1, %r15d addq 1040(%rsp), %r14 movq 8(%rsp), %rax # 8-byte Reload adcq 1048(%rsp), %rax movq %rax, 8(%rsp) # 8-byte Spill movq 48(%rsp), %rax # 8-byte Reload adcq 1056(%rsp), %rax movq %rax, %r12 movq 64(%rsp), %rax # 8-byte Reload adcq 1064(%rsp), %rax movq %rax, 64(%rsp) # 8-byte Spill adcq 1072(%rsp), %r13 movq %r13, 72(%rsp) # 8-byte Spill adcq 1080(%rsp), %rbx movq %rbx, 88(%rsp) # 8-byte Spill adcq 1088(%rsp), %rbp movq 96(%rsp), %rax # 8-byte Reload adcq 1096(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill adcq 1104(%rsp), %r15 movq %r15, 56(%rsp) # 8-byte Spill sbbq %r15, %r15 movq %r14, %rdx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 968(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 andl $1, %r15d addq 968(%rsp), %r14 movq 8(%rsp), %r13 # 8-byte Reload adcq 976(%rsp), %r13 adcq 984(%rsp), %r12 movq %r12, 48(%rsp) # 8-byte Spill movq 64(%rsp), %r14 # 8-byte Reload adcq 992(%rsp), %r14 movq 72(%rsp), %rbx # 8-byte Reload adcq 1000(%rsp), %rbx movq 88(%rsp), %rax # 8-byte Reload adcq 1008(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill adcq 1016(%rsp), %rbp movq %rbp, %r12 movq 96(%rsp), %rbp # 8-byte Reload adcq 1024(%rsp), %rbp movq 56(%rsp), %rax # 8-byte Reload adcq 1032(%rsp), %rax movq %rax, 56(%rsp) # 8-byte Spill adcq $0, %r15 movq 16(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdx leaq 896(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq %r13, %rcx addq 896(%rsp), %rcx movq 48(%rsp), %r13 # 8-byte Reload adcq 904(%rsp), %r13 adcq 912(%rsp), %r14 adcq 920(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill movq 88(%rsp), %rax # 8-byte Reload adcq 928(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill adcq 936(%rsp), %r12 movq %r12, 80(%rsp) # 8-byte Spill adcq 944(%rsp), %rbp movq %rbp, 96(%rsp) # 8-byte Spill movq 56(%rsp), %r12 # 8-byte Reload adcq 952(%rsp), %r12 adcq 960(%rsp), %r15 sbbq %rbx, %rbx movq %rcx, %rdx movq %rcx, %rbp imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 824(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 andl $1, %ebx addq 824(%rsp), %rbp adcq 832(%rsp), %r13 movq %r13, 48(%rsp) # 8-byte Spill adcq 840(%rsp), %r14 movq %r14, 64(%rsp) # 8-byte Spill movq 72(%rsp), %r13 # 8-byte Reload adcq 848(%rsp), %r13 movq 88(%rsp), %rbp # 8-byte Reload adcq 856(%rsp), %rbp movq 80(%rsp), %r14 # 8-byte Reload adcq 864(%rsp), %r14 movq 96(%rsp), %rax # 8-byte Reload adcq 872(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill adcq 880(%rsp), %r12 adcq 888(%rsp), %r15 adcq $0, %rbx movq 16(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdx leaq 752(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 48(%rsp), %rax # 8-byte Reload addq 752(%rsp), %rax movq 64(%rsp), %rcx # 8-byte Reload adcq 760(%rsp), %rcx movq %rcx, 64(%rsp) # 8-byte Spill adcq 768(%rsp), %r13 movq %r13, 72(%rsp) # 8-byte Spill adcq 776(%rsp), %rbp movq %rbp, 88(%rsp) # 8-byte Spill adcq 784(%rsp), %r14 movq %r14, 80(%rsp) # 8-byte Spill movq 96(%rsp), %rbp # 8-byte Reload adcq 792(%rsp), %rbp adcq 800(%rsp), %r12 adcq 808(%rsp), %r15 adcq 816(%rsp), %rbx movq %rbx, 48(%rsp) # 8-byte Spill sbbq %r13, %r13 movq %rax, %rdx movq %rax, %rbx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 680(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq %r13, %rax andl $1, %eax addq 680(%rsp), %rbx movq 64(%rsp), %r14 # 8-byte Reload adcq 688(%rsp), %r14 movq 72(%rsp), %rcx # 8-byte Reload adcq 696(%rsp), %rcx movq %rcx, 72(%rsp) # 8-byte Spill movq 88(%rsp), %r13 # 8-byte Reload adcq 704(%rsp), %r13 movq 80(%rsp), %rbx # 8-byte Reload adcq 712(%rsp), %rbx adcq 720(%rsp), %rbp movq %rbp, 96(%rsp) # 8-byte Spill movq %r12, %rbp adcq 728(%rsp), %rbp adcq 736(%rsp), %r15 movq 48(%rsp), %r12 # 8-byte Reload adcq 744(%rsp), %r12 adcq $0, %rax movq %rax, 64(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 32(%rax), %rdx leaq 608(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq %r14, %rax addq 608(%rsp), %rax movq 72(%rsp), %r14 # 8-byte Reload adcq 616(%rsp), %r14 adcq 624(%rsp), %r13 movq %r13, 88(%rsp) # 8-byte Spill adcq 632(%rsp), %rbx movq %rbx, %r13 movq 96(%rsp), %rcx # 8-byte Reload adcq 640(%rsp), %rcx movq %rcx, 96(%rsp) # 8-byte Spill adcq 648(%rsp), %rbp movq %rbp, 56(%rsp) # 8-byte Spill adcq 656(%rsp), %r15 adcq 664(%rsp), %r12 movq %r12, 48(%rsp) # 8-byte Spill movq 64(%rsp), %rcx # 8-byte Reload adcq 672(%rsp), %rcx movq %rcx, 64(%rsp) # 8-byte Spill sbbq %rbp, %rbp movq %rax, %rdx movq %rax, %rbx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 536(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq %rbp, %rax andl $1, %eax addq 536(%rsp), %rbx adcq 544(%rsp), %r14 movq %r14, 72(%rsp) # 8-byte Spill movq 88(%rsp), %rbx # 8-byte Reload adcq 552(%rsp), %rbx adcq 560(%rsp), %r13 movq 96(%rsp), %rbp # 8-byte Reload adcq 568(%rsp), %rbp movq 56(%rsp), %r12 # 8-byte Reload adcq 576(%rsp), %r12 adcq 584(%rsp), %r15 movq 48(%rsp), %rcx # 8-byte Reload adcq 592(%rsp), %rcx movq %rcx, 48(%rsp) # 8-byte Spill movq 64(%rsp), %r14 # 8-byte Reload adcq 600(%rsp), %r14 adcq $0, %rax movq %rax, 88(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 40(%rax), %rdx leaq 464(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 72(%rsp), %rax # 8-byte Reload addq 464(%rsp), %rax adcq 472(%rsp), %rbx adcq 480(%rsp), %r13 movq %r13, 80(%rsp) # 8-byte Spill adcq 488(%rsp), %rbp movq %rbp, 96(%rsp) # 8-byte Spill adcq 496(%rsp), %r12 adcq 504(%rsp), %r15 movq %r15, 72(%rsp) # 8-byte Spill movq 48(%rsp), %r15 # 8-byte Reload adcq 512(%rsp), %r15 adcq 520(%rsp), %r14 movq %r14, 64(%rsp) # 8-byte Spill movq 88(%rsp), %r14 # 8-byte Reload adcq 528(%rsp), %r14 sbbq %r13, %r13 movq %rax, %rdx movq %rax, %rbp imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 392(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq %r13, %rax andl $1, %eax addq 392(%rsp), %rbp adcq 400(%rsp), %rbx movq %rbx, 88(%rsp) # 8-byte Spill movq 80(%rsp), %rbp # 8-byte Reload adcq 408(%rsp), %rbp movq 96(%rsp), %rbx # 8-byte Reload adcq 416(%rsp), %rbx adcq 424(%rsp), %r12 movq 72(%rsp), %r13 # 8-byte Reload adcq 432(%rsp), %r13 adcq 440(%rsp), %r15 movq %r15, 48(%rsp) # 8-byte Spill movq 64(%rsp), %r15 # 8-byte Reload adcq 448(%rsp), %r15 adcq 456(%rsp), %r14 adcq $0, %rax movq %rax, 72(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 48(%rax), %rdx leaq 320(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 88(%rsp), %rax # 8-byte Reload addq 320(%rsp), %rax adcq 328(%rsp), %rbp movq %rbp, 80(%rsp) # 8-byte Spill adcq 336(%rsp), %rbx movq %rbx, 96(%rsp) # 8-byte Spill movq %r12, %rbp adcq 344(%rsp), %rbp adcq 352(%rsp), %r13 movq 48(%rsp), %r12 # 8-byte Reload adcq 360(%rsp), %r12 adcq 368(%rsp), %r15 movq %r15, 64(%rsp) # 8-byte Spill adcq 376(%rsp), %r14 movq %r14, 88(%rsp) # 8-byte Spill movq 72(%rsp), %rcx # 8-byte Reload adcq 384(%rsp), %rcx movq %rcx, 72(%rsp) # 8-byte Spill sbbq %r15, %r15 movq %rax, %rdx movq %rax, %rbx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 248(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 andl $1, %r15d addq 248(%rsp), %rbx movq 80(%rsp), %rax # 8-byte Reload adcq 256(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 96(%rsp), %r14 # 8-byte Reload adcq 264(%rsp), %r14 adcq 272(%rsp), %rbp movq %rbp, 56(%rsp) # 8-byte Spill movq %r13, %rbx adcq 280(%rsp), %rbx movq %r12, %rbp adcq 288(%rsp), %rbp movq 64(%rsp), %r13 # 8-byte Reload adcq 296(%rsp), %r13 movq 88(%rsp), %rax # 8-byte Reload adcq 304(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 72(%rsp), %r12 # 8-byte Reload adcq 312(%rsp), %r12 adcq $0, %r15 movq 16(%rsp), %rax # 8-byte Reload movq 56(%rax), %rdx leaq 176(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 80(%rsp), %rax # 8-byte Reload addq 176(%rsp), %rax adcq 184(%rsp), %r14 movq %r14, 96(%rsp) # 8-byte Spill movq 56(%rsp), %rcx # 8-byte Reload adcq 192(%rsp), %rcx movq %rcx, 56(%rsp) # 8-byte Spill adcq 200(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill adcq 208(%rsp), %rbp adcq 216(%rsp), %r13 movq %r13, 64(%rsp) # 8-byte Spill movq 88(%rsp), %r14 # 8-byte Reload adcq 224(%rsp), %r14 adcq 232(%rsp), %r12 adcq 240(%rsp), %r15 sbbq %rbx, %rbx movq 32(%rsp), %rdx # 8-byte Reload imulq %rax, %rdx movq %rax, %r13 leaq 104(%rsp), %rdi movq 40(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 andl $1, %ebx addq 104(%rsp), %r13 movq 96(%rsp), %rcx # 8-byte Reload adcq 112(%rsp), %rcx movq 56(%rsp), %rdx # 8-byte Reload adcq 120(%rsp), %rdx movq 72(%rsp), %rsi # 8-byte Reload adcq 128(%rsp), %rsi movq %rbp, %rdi adcq 136(%rsp), %rdi movq %rdi, 48(%rsp) # 8-byte Spill movq 64(%rsp), %r8 # 8-byte Reload adcq 144(%rsp), %r8 movq %r8, 64(%rsp) # 8-byte Spill movq %r14, %r9 adcq 152(%rsp), %r9 movq %r9, 88(%rsp) # 8-byte Spill adcq 160(%rsp), %r12 adcq 168(%rsp), %r15 adcq $0, %rbx movq %rcx, %rax movq %rcx, %r11 movq 40(%rsp), %rbp # 8-byte Reload subq (%rbp), %rax movq %rdx, %rcx movq %rdx, %r14 sbbq 8(%rbp), %rcx movq %rsi, %rdx movq %rsi, %r13 sbbq 16(%rbp), %rdx movq %rdi, %rsi sbbq 24(%rbp), %rsi movq %r8, %rdi sbbq 32(%rbp), %rdi movq %r9, %r10 sbbq 40(%rbp), %r10 movq %r12, %r8 sbbq 48(%rbp), %r8 movq %r15, %r9 sbbq 56(%rbp), %r9 sbbq $0, %rbx andl $1, %ebx cmovneq %r15, %r9 testb %bl, %bl cmovneq %r11, %rax movq (%rsp), %rbx # 8-byte Reload movq %rax, (%rbx) cmovneq %r14, %rcx movq %rcx, 8(%rbx) cmovneq %r13, %rdx movq %rdx, 16(%rbx) cmovneq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 24(%rbx) cmovneq 64(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rbx) cmovneq 88(%rsp), %r10 # 8-byte Folded Reload movq %r10, 40(%rbx) cmovneq %r12, %r8 movq %r8, 48(%rbx) movq %r9, 56(%rbx) addq $1256, %rsp # imm = 0x4E8 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end114: .size mcl_fp_mont8L, .Lfunc_end114-mcl_fp_mont8L .globl mcl_fp_montNF8L .align 16, 0x90 .type mcl_fp_montNF8L,@function mcl_fp_montNF8L: # @mcl_fp_montNF8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $1240, %rsp # imm = 0x4D8 movq %rcx, 56(%rsp) # 8-byte Spill movq %rdx, 16(%rsp) # 8-byte Spill movq %rsi, 24(%rsp) # 8-byte Spill movq %rdi, (%rsp) # 8-byte Spill movq -8(%rcx), %rbx movq %rbx, 32(%rsp) # 8-byte Spill movq (%rdx), %rdx leaq 1168(%rsp), %rdi callq .LmulPv512x64 movq 1168(%rsp), %r15 movq 1176(%rsp), %r12 movq %r15, %rdx imulq %rbx, %rdx movq 1232(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 1224(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 1216(%rsp), %r13 movq 1208(%rsp), %rax movq %rax, 72(%rsp) # 8-byte Spill movq 1200(%rsp), %r14 movq 1192(%rsp), %rbp movq 1184(%rsp), %rbx leaq 1096(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 1096(%rsp), %r15 adcq 1104(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill adcq 1112(%rsp), %rbx adcq 1120(%rsp), %rbp adcq 1128(%rsp), %r14 movq %r14, %r12 movq 72(%rsp), %r14 # 8-byte Reload adcq 1136(%rsp), %r14 adcq 1144(%rsp), %r13 movq 80(%rsp), %rax # 8-byte Reload adcq 1152(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 48(%rsp), %rax # 8-byte Reload adcq 1160(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdx leaq 1024(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 1088(%rsp), %r15 movq 64(%rsp), %rax # 8-byte Reload addq 1024(%rsp), %rax adcq 1032(%rsp), %rbx movq %rbx, 8(%rsp) # 8-byte Spill movq %rbp, %rbx adcq 1040(%rsp), %rbx adcq 1048(%rsp), %r12 adcq 1056(%rsp), %r14 movq %r14, 72(%rsp) # 8-byte Spill movq %r13, %rbp adcq 1064(%rsp), %rbp movq 80(%rsp), %rcx # 8-byte Reload adcq 1072(%rsp), %rcx movq %rcx, 80(%rsp) # 8-byte Spill movq 48(%rsp), %r14 # 8-byte Reload adcq 1080(%rsp), %r14 adcq $0, %r15 movq %rax, %rdx movq %rax, %r13 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 952(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 952(%rsp), %r13 movq 8(%rsp), %rax # 8-byte Reload adcq 960(%rsp), %rax movq %rax, 8(%rsp) # 8-byte Spill adcq 968(%rsp), %rbx movq %rbx, 64(%rsp) # 8-byte Spill movq %r12, %rbx adcq 976(%rsp), %rbx movq 72(%rsp), %r12 # 8-byte Reload adcq 984(%rsp), %r12 adcq 992(%rsp), %rbp movq %rbp, 40(%rsp) # 8-byte Spill movq 80(%rsp), %r13 # 8-byte Reload adcq 1000(%rsp), %r13 movq %r14, %rbp adcq 1008(%rsp), %rbp adcq 1016(%rsp), %r15 movq 16(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdx leaq 880(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 944(%rsp), %r14 movq 8(%rsp), %rax # 8-byte Reload addq 880(%rsp), %rax movq 64(%rsp), %rcx # 8-byte Reload adcq 888(%rsp), %rcx movq %rcx, 64(%rsp) # 8-byte Spill adcq 896(%rsp), %rbx adcq 904(%rsp), %r12 movq %r12, 72(%rsp) # 8-byte Spill movq 40(%rsp), %rcx # 8-byte Reload adcq 912(%rsp), %rcx movq %rcx, 40(%rsp) # 8-byte Spill adcq 920(%rsp), %r13 movq %r13, 80(%rsp) # 8-byte Spill adcq 928(%rsp), %rbp movq %rbp, 48(%rsp) # 8-byte Spill adcq 936(%rsp), %r15 adcq $0, %r14 movq %rax, %rdx movq %rax, %rbp imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 808(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 808(%rsp), %rbp movq 64(%rsp), %r13 # 8-byte Reload adcq 816(%rsp), %r13 movq %rbx, %r12 adcq 824(%rsp), %r12 movq 72(%rsp), %rbx # 8-byte Reload adcq 832(%rsp), %rbx movq 40(%rsp), %rbp # 8-byte Reload adcq 840(%rsp), %rbp movq 80(%rsp), %rax # 8-byte Reload adcq 848(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 48(%rsp), %rax # 8-byte Reload adcq 856(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill adcq 864(%rsp), %r15 adcq 872(%rsp), %r14 movq 16(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdx leaq 736(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 800(%rsp), %rax movq %r13, %rcx addq 736(%rsp), %rcx adcq 744(%rsp), %r12 movq %r12, 40(%rsp) # 8-byte Spill adcq 752(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill adcq 760(%rsp), %rbp movq %rbp, %r13 movq 80(%rsp), %rbp # 8-byte Reload adcq 768(%rsp), %rbp movq 48(%rsp), %rbx # 8-byte Reload adcq 776(%rsp), %rbx adcq 784(%rsp), %r15 adcq 792(%rsp), %r14 adcq $0, %rax movq %rax, 64(%rsp) # 8-byte Spill movq %rcx, %rdx movq %rcx, %r12 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 664(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 664(%rsp), %r12 movq 40(%rsp), %rax # 8-byte Reload adcq 672(%rsp), %rax movq %rax, 40(%rsp) # 8-byte Spill movq 72(%rsp), %rax # 8-byte Reload adcq 680(%rsp), %rax movq %rax, 72(%rsp) # 8-byte Spill adcq 688(%rsp), %r13 adcq 696(%rsp), %rbp movq %rbp, 80(%rsp) # 8-byte Spill adcq 704(%rsp), %rbx adcq 712(%rsp), %r15 adcq 720(%rsp), %r14 movq 64(%rsp), %r12 # 8-byte Reload adcq 728(%rsp), %r12 movq 16(%rsp), %rax # 8-byte Reload movq 32(%rax), %rdx leaq 592(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 656(%rsp), %rcx movq 40(%rsp), %rax # 8-byte Reload addq 592(%rsp), %rax movq 72(%rsp), %rbp # 8-byte Reload adcq 600(%rsp), %rbp adcq 608(%rsp), %r13 movq %r13, 40(%rsp) # 8-byte Spill movq 80(%rsp), %r13 # 8-byte Reload adcq 616(%rsp), %r13 adcq 624(%rsp), %rbx adcq 632(%rsp), %r15 adcq 640(%rsp), %r14 adcq 648(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill adcq $0, %rcx movq %rcx, 80(%rsp) # 8-byte Spill movq %rax, %rdx movq %rax, %r12 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 520(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 520(%rsp), %r12 adcq 528(%rsp), %rbp movq %rbp, 72(%rsp) # 8-byte Spill movq 40(%rsp), %r12 # 8-byte Reload adcq 536(%rsp), %r12 movq %r13, %rbp adcq 544(%rsp), %rbp adcq 552(%rsp), %rbx adcq 560(%rsp), %r15 adcq 568(%rsp), %r14 movq 64(%rsp), %r13 # 8-byte Reload adcq 576(%rsp), %r13 movq 80(%rsp), %rax # 8-byte Reload adcq 584(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 40(%rax), %rdx leaq 448(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 512(%rsp), %rcx movq 72(%rsp), %rax # 8-byte Reload addq 448(%rsp), %rax adcq 456(%rsp), %r12 movq %r12, 40(%rsp) # 8-byte Spill adcq 464(%rsp), %rbp adcq 472(%rsp), %rbx adcq 480(%rsp), %r15 adcq 488(%rsp), %r14 adcq 496(%rsp), %r13 movq %r13, 64(%rsp) # 8-byte Spill movq 80(%rsp), %r13 # 8-byte Reload adcq 504(%rsp), %r13 adcq $0, %rcx movq %rcx, 72(%rsp) # 8-byte Spill movq %rax, %rdx movq %rax, %r12 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 376(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 376(%rsp), %r12 movq 40(%rsp), %rax # 8-byte Reload adcq 384(%rsp), %rax movq %rax, 40(%rsp) # 8-byte Spill adcq 392(%rsp), %rbp adcq 400(%rsp), %rbx adcq 408(%rsp), %r15 adcq 416(%rsp), %r14 movq 64(%rsp), %r12 # 8-byte Reload adcq 424(%rsp), %r12 adcq 432(%rsp), %r13 movq 72(%rsp), %rax # 8-byte Reload adcq 440(%rsp), %rax movq %rax, 72(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 48(%rax), %rdx leaq 304(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 368(%rsp), %rcx movq 40(%rsp), %rax # 8-byte Reload addq 304(%rsp), %rax adcq 312(%rsp), %rbp movq %rbp, 80(%rsp) # 8-byte Spill adcq 320(%rsp), %rbx adcq 328(%rsp), %r15 adcq 336(%rsp), %r14 adcq 344(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill adcq 352(%rsp), %r13 movq 72(%rsp), %rbp # 8-byte Reload adcq 360(%rsp), %rbp adcq $0, %rcx movq %rcx, 48(%rsp) # 8-byte Spill movq %rax, %rdx movq %rax, %r12 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 232(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 232(%rsp), %r12 movq 80(%rsp), %rax # 8-byte Reload adcq 240(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill adcq 248(%rsp), %rbx adcq 256(%rsp), %r15 adcq 264(%rsp), %r14 movq 64(%rsp), %r12 # 8-byte Reload adcq 272(%rsp), %r12 adcq 280(%rsp), %r13 adcq 288(%rsp), %rbp movq %rbp, 72(%rsp) # 8-byte Spill movq 48(%rsp), %rbp # 8-byte Reload adcq 296(%rsp), %rbp movq 16(%rsp), %rax # 8-byte Reload movq 56(%rax), %rdx leaq 160(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 movq 224(%rsp), %rcx movq 80(%rsp), %rax # 8-byte Reload addq 160(%rsp), %rax adcq 168(%rsp), %rbx movq %rbx, 48(%rsp) # 8-byte Spill adcq 176(%rsp), %r15 adcq 184(%rsp), %r14 adcq 192(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill adcq 200(%rsp), %r13 movq 72(%rsp), %rbx # 8-byte Reload adcq 208(%rsp), %rbx adcq 216(%rsp), %rbp movq %rbp, %r12 adcq $0, %rcx movq %rcx, 80(%rsp) # 8-byte Spill movq 32(%rsp), %rdx # 8-byte Reload imulq %rax, %rdx movq %rax, %rbp leaq 88(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 88(%rsp), %rbp movq 48(%rsp), %r11 # 8-byte Reload adcq 96(%rsp), %r11 adcq 104(%rsp), %r15 adcq 112(%rsp), %r14 movq 64(%rsp), %rsi # 8-byte Reload adcq 120(%rsp), %rsi movq %rsi, 64(%rsp) # 8-byte Spill adcq 128(%rsp), %r13 adcq 136(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill adcq 144(%rsp), %r12 movq 80(%rsp), %r8 # 8-byte Reload adcq 152(%rsp), %r8 movq %r11, %rax movq 56(%rsp), %rbp # 8-byte Reload subq (%rbp), %rax movq %r15, %rcx sbbq 8(%rbp), %rcx movq %r14, %rdx sbbq 16(%rbp), %rdx sbbq 24(%rbp), %rsi movq %r13, %rdi sbbq 32(%rbp), %rdi movq %rbx, %r9 sbbq 40(%rbp), %r9 movq %r12, %r10 sbbq 48(%rbp), %r10 movq %rbp, %rbx movq %r8, %rbp sbbq 56(%rbx), %rbp testq %rbp, %rbp cmovsq %r11, %rax movq (%rsp), %rbx # 8-byte Reload movq %rax, (%rbx) cmovsq %r15, %rcx movq %rcx, 8(%rbx) cmovsq %r14, %rdx movq %rdx, 16(%rbx) cmovsq 64(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 24(%rbx) cmovsq %r13, %rdi movq %rdi, 32(%rbx) cmovsq 72(%rsp), %r9 # 8-byte Folded Reload movq %r9, 40(%rbx) cmovsq %r12, %r10 movq %r10, 48(%rbx) cmovsq %r8, %rbp movq %rbp, 56(%rbx) addq $1240, %rsp # imm = 0x4D8 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end115: .size mcl_fp_montNF8L, .Lfunc_end115-mcl_fp_montNF8L .globl mcl_fp_montRed8L .align 16, 0x90 .type mcl_fp_montRed8L,@function mcl_fp_montRed8L: # @mcl_fp_montRed8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $776, %rsp # imm = 0x308 movq %rdx, %rax movq %rax, 112(%rsp) # 8-byte Spill movq %rdi, 72(%rsp) # 8-byte Spill movq -8(%rax), %rcx movq %rcx, 128(%rsp) # 8-byte Spill movq (%rsi), %r15 movq 8(%rsi), %rdx movq %rdx, 184(%rsp) # 8-byte Spill movq %r15, %rdx imulq %rcx, %rdx movq 120(%rsi), %rcx movq %rcx, 96(%rsp) # 8-byte Spill movq 112(%rsi), %rcx movq %rcx, 136(%rsp) # 8-byte Spill movq 104(%rsi), %rcx movq %rcx, 120(%rsp) # 8-byte Spill movq 96(%rsi), %rcx movq %rcx, 168(%rsp) # 8-byte Spill movq 88(%rsi), %rcx movq %rcx, 176(%rsp) # 8-byte Spill movq 80(%rsi), %rcx movq %rcx, 160(%rsp) # 8-byte Spill movq 72(%rsi), %rcx movq %rcx, 192(%rsp) # 8-byte Spill movq 64(%rsi), %r13 movq 56(%rsi), %rcx movq %rcx, 144(%rsp) # 8-byte Spill movq 48(%rsi), %r14 movq 40(%rsi), %rcx movq %rcx, 152(%rsp) # 8-byte Spill movq 32(%rsi), %r12 movq 24(%rsi), %rbx movq 16(%rsi), %rbp movq %rax, %rcx movq (%rcx), %rax movq %rax, 16(%rsp) # 8-byte Spill movq 56(%rcx), %rax movq %rax, 64(%rsp) # 8-byte Spill movq 48(%rcx), %rax movq %rax, 56(%rsp) # 8-byte Spill movq 40(%rcx), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 32(%rcx), %rax movq %rax, 40(%rsp) # 8-byte Spill movq 24(%rcx), %rax movq %rax, 32(%rsp) # 8-byte Spill movq 16(%rcx), %rax movq %rax, 24(%rsp) # 8-byte Spill movq 8(%rcx), %rax movq %rax, 8(%rsp) # 8-byte Spill movq %rcx, %rsi leaq 704(%rsp), %rdi callq .LmulPv512x64 addq 704(%rsp), %r15 movq 184(%rsp), %rcx # 8-byte Reload adcq 712(%rsp), %rcx adcq 720(%rsp), %rbp movq %rbp, 80(%rsp) # 8-byte Spill adcq 728(%rsp), %rbx movq %rbx, 88(%rsp) # 8-byte Spill adcq 736(%rsp), %r12 movq %r12, 104(%rsp) # 8-byte Spill movq 152(%rsp), %rax # 8-byte Reload adcq 744(%rsp), %rax movq %rax, 152(%rsp) # 8-byte Spill adcq 752(%rsp), %r14 movq %r14, %r12 movq 144(%rsp), %rax # 8-byte Reload adcq 760(%rsp), %rax movq %rax, 144(%rsp) # 8-byte Spill adcq 768(%rsp), %r13 movq %r13, 184(%rsp) # 8-byte Spill adcq $0, 192(%rsp) # 8-byte Folded Spill movq 160(%rsp), %r15 # 8-byte Reload adcq $0, %r15 adcq $0, 176(%rsp) # 8-byte Folded Spill adcq $0, 168(%rsp) # 8-byte Folded Spill adcq $0, 120(%rsp) # 8-byte Folded Spill movq 136(%rsp), %r13 # 8-byte Reload adcq $0, %r13 movq 96(%rsp), %r14 # 8-byte Reload adcq $0, %r14 sbbq %rbx, %rbx movq %rcx, %rbp movq %rbp, %rdx imulq 128(%rsp), %rdx # 8-byte Folded Reload leaq 632(%rsp), %rdi movq 112(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 andl $1, %ebx movq %rbx, %rax addq 632(%rsp), %rbp movq 80(%rsp), %rsi # 8-byte Reload adcq 640(%rsp), %rsi movq 88(%rsp), %rcx # 8-byte Reload adcq 648(%rsp), %rcx movq %rcx, 88(%rsp) # 8-byte Spill movq 104(%rsp), %rcx # 8-byte Reload adcq 656(%rsp), %rcx movq %rcx, 104(%rsp) # 8-byte Spill movq 152(%rsp), %rcx # 8-byte Reload adcq 664(%rsp), %rcx movq %rcx, 152(%rsp) # 8-byte Spill adcq 672(%rsp), %r12 movq 144(%rsp), %rcx # 8-byte Reload adcq 680(%rsp), %rcx movq %rcx, 144(%rsp) # 8-byte Spill movq 184(%rsp), %rcx # 8-byte Reload adcq 688(%rsp), %rcx movq %rcx, 184(%rsp) # 8-byte Spill movq 192(%rsp), %rcx # 8-byte Reload adcq 696(%rsp), %rcx movq %rcx, 192(%rsp) # 8-byte Spill adcq $0, %r15 movq %r15, 160(%rsp) # 8-byte Spill movq 176(%rsp), %rbx # 8-byte Reload adcq $0, %rbx movq 168(%rsp), %r15 # 8-byte Reload adcq $0, %r15 adcq $0, 120(%rsp) # 8-byte Folded Spill adcq $0, %r13 movq %r13, 136(%rsp) # 8-byte Spill adcq $0, %r14 movq %r14, 96(%rsp) # 8-byte Spill movq %rax, %rbp adcq $0, %rbp movq %rsi, %rdx movq %rsi, %r14 imulq 128(%rsp), %rdx # 8-byte Folded Reload leaq 560(%rsp), %rdi movq 112(%rsp), %r13 # 8-byte Reload movq %r13, %rsi callq .LmulPv512x64 addq 560(%rsp), %r14 movq 88(%rsp), %rcx # 8-byte Reload adcq 568(%rsp), %rcx movq 104(%rsp), %rax # 8-byte Reload adcq 576(%rsp), %rax movq %rax, 104(%rsp) # 8-byte Spill movq 152(%rsp), %rax # 8-byte Reload adcq 584(%rsp), %rax movq %rax, 152(%rsp) # 8-byte Spill adcq 592(%rsp), %r12 movq %r12, 88(%rsp) # 8-byte Spill movq 144(%rsp), %r14 # 8-byte Reload adcq 600(%rsp), %r14 movq 184(%rsp), %rax # 8-byte Reload adcq 608(%rsp), %rax movq %rax, 184(%rsp) # 8-byte Spill movq 192(%rsp), %rax # 8-byte Reload adcq 616(%rsp), %rax movq %rax, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rax # 8-byte Reload adcq 624(%rsp), %rax movq %rax, 160(%rsp) # 8-byte Spill adcq $0, %rbx movq %rbx, 176(%rsp) # 8-byte Spill adcq $0, %r15 movq %r15, 168(%rsp) # 8-byte Spill movq 120(%rsp), %rbx # 8-byte Reload adcq $0, %rbx movq 136(%rsp), %r15 # 8-byte Reload adcq $0, %r15 adcq $0, 96(%rsp) # 8-byte Folded Spill adcq $0, %rbp movq %rbp, 80(%rsp) # 8-byte Spill movq %rcx, %rbp movq %rbp, %rdx movq 128(%rsp), %r12 # 8-byte Reload imulq %r12, %rdx leaq 488(%rsp), %rdi movq %r13, %rsi callq .LmulPv512x64 addq 488(%rsp), %rbp movq 104(%rsp), %rax # 8-byte Reload adcq 496(%rsp), %rax movq 152(%rsp), %rbp # 8-byte Reload adcq 504(%rsp), %rbp movq 88(%rsp), %rcx # 8-byte Reload adcq 512(%rsp), %rcx movq %rcx, 88(%rsp) # 8-byte Spill adcq 520(%rsp), %r14 movq 184(%rsp), %rcx # 8-byte Reload adcq 528(%rsp), %rcx movq %rcx, 184(%rsp) # 8-byte Spill movq 192(%rsp), %rcx # 8-byte Reload adcq 536(%rsp), %rcx movq %rcx, 192(%rsp) # 8-byte Spill movq 160(%rsp), %r13 # 8-byte Reload adcq 544(%rsp), %r13 movq 176(%rsp), %rcx # 8-byte Reload adcq 552(%rsp), %rcx movq %rcx, 176(%rsp) # 8-byte Spill adcq $0, 168(%rsp) # 8-byte Folded Spill adcq $0, %rbx movq %rbx, 120(%rsp) # 8-byte Spill movq %r15, %rbx adcq $0, %rbx adcq $0, 96(%rsp) # 8-byte Folded Spill adcq $0, 80(%rsp) # 8-byte Folded Spill movq %rax, %rdx movq %rax, %r15 imulq %r12, %rdx leaq 416(%rsp), %rdi movq 112(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 416(%rsp), %r15 adcq 424(%rsp), %rbp movq %rbp, %rax movq 88(%rsp), %rcx # 8-byte Reload adcq 432(%rsp), %rcx movq %rcx, 88(%rsp) # 8-byte Spill movq %r14, %r12 adcq 440(%rsp), %r12 movq 184(%rsp), %r14 # 8-byte Reload adcq 448(%rsp), %r14 movq 192(%rsp), %rbp # 8-byte Reload adcq 456(%rsp), %rbp adcq 464(%rsp), %r13 movq 176(%rsp), %rcx # 8-byte Reload adcq 472(%rsp), %rcx movq %rcx, 176(%rsp) # 8-byte Spill movq 168(%rsp), %rcx # 8-byte Reload adcq 480(%rsp), %rcx movq %rcx, 168(%rsp) # 8-byte Spill adcq $0, 120(%rsp) # 8-byte Folded Spill adcq $0, %rbx movq %rbx, 136(%rsp) # 8-byte Spill movq 96(%rsp), %r15 # 8-byte Reload adcq $0, %r15 adcq $0, 80(%rsp) # 8-byte Folded Spill movq %rax, %rbx movq %rbx, %rdx imulq 128(%rsp), %rdx # 8-byte Folded Reload leaq 344(%rsp), %rdi movq 112(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 344(%rsp), %rbx movq 88(%rsp), %rax # 8-byte Reload adcq 352(%rsp), %rax adcq 360(%rsp), %r12 movq %r12, 144(%rsp) # 8-byte Spill adcq 368(%rsp), %r14 movq %r14, 184(%rsp) # 8-byte Spill adcq 376(%rsp), %rbp movq %rbp, 192(%rsp) # 8-byte Spill adcq 384(%rsp), %r13 movq %r13, 160(%rsp) # 8-byte Spill movq 176(%rsp), %r13 # 8-byte Reload adcq 392(%rsp), %r13 movq 168(%rsp), %r12 # 8-byte Reload adcq 400(%rsp), %r12 movq 120(%rsp), %r14 # 8-byte Reload adcq 408(%rsp), %r14 movq 136(%rsp), %rbp # 8-byte Reload adcq $0, %rbp movq %r15, %rbx adcq $0, %rbx adcq $0, 80(%rsp) # 8-byte Folded Spill movq %rax, %rdx movq %rax, %r15 imulq 128(%rsp), %rdx # 8-byte Folded Reload leaq 272(%rsp), %rdi movq 112(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 272(%rsp), %r15 movq 144(%rsp), %rcx # 8-byte Reload adcq 280(%rsp), %rcx movq 184(%rsp), %rax # 8-byte Reload adcq 288(%rsp), %rax movq %rax, 184(%rsp) # 8-byte Spill movq 192(%rsp), %rax # 8-byte Reload adcq 296(%rsp), %rax movq %rax, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rax # 8-byte Reload adcq 304(%rsp), %rax movq %rax, 160(%rsp) # 8-byte Spill adcq 312(%rsp), %r13 movq %r13, 176(%rsp) # 8-byte Spill adcq 320(%rsp), %r12 movq %r12, 168(%rsp) # 8-byte Spill adcq 328(%rsp), %r14 movq %r14, %r13 adcq 336(%rsp), %rbp movq %rbp, %r12 adcq $0, %rbx movq %rbx, %r14 movq 80(%rsp), %r15 # 8-byte Reload adcq $0, %r15 movq 128(%rsp), %rdx # 8-byte Reload movq %rcx, %rbx imulq %rbx, %rdx leaq 200(%rsp), %rdi movq 112(%rsp), %rsi # 8-byte Reload callq .LmulPv512x64 addq 200(%rsp), %rbx movq 184(%rsp), %rax # 8-byte Reload adcq 208(%rsp), %rax movq %rax, 184(%rsp) # 8-byte Spill movq 192(%rsp), %r8 # 8-byte Reload adcq 216(%rsp), %r8 movq %r8, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rdx # 8-byte Reload adcq 224(%rsp), %rdx movq 176(%rsp), %rsi # 8-byte Reload adcq 232(%rsp), %rsi movq 168(%rsp), %rdi # 8-byte Reload adcq 240(%rsp), %rdi movq %r13, %rbp adcq 248(%rsp), %rbp movq %r12, %rbx adcq 256(%rsp), %rbx movq %rbx, 136(%rsp) # 8-byte Spill movq %r14, %r9 adcq 264(%rsp), %r9 adcq $0, %r15 movq %r15, %r10 subq 16(%rsp), %rax # 8-byte Folded Reload movq %r8, %rcx sbbq 8(%rsp), %rcx # 8-byte Folded Reload movq %rdx, %r13 sbbq 24(%rsp), %r13 # 8-byte Folded Reload movq %rsi, %r12 sbbq 32(%rsp), %r12 # 8-byte Folded Reload movq %rdi, %r14 sbbq 40(%rsp), %r14 # 8-byte Folded Reload movq %rbp, %r11 sbbq 48(%rsp), %r11 # 8-byte Folded Reload movq %rbx, %r8 sbbq 56(%rsp), %r8 # 8-byte Folded Reload movq %r9, %r15 sbbq 64(%rsp), %r9 # 8-byte Folded Reload sbbq $0, %r10 andl $1, %r10d cmovneq %r15, %r9 testb %r10b, %r10b cmovneq 184(%rsp), %rax # 8-byte Folded Reload movq 72(%rsp), %rbx # 8-byte Reload movq %rax, (%rbx) cmovneq 192(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 8(%rbx) cmovneq %rdx, %r13 movq %r13, 16(%rbx) cmovneq %rsi, %r12 movq %r12, 24(%rbx) cmovneq %rdi, %r14 movq %r14, 32(%rbx) cmovneq %rbp, %r11 movq %r11, 40(%rbx) cmovneq 136(%rsp), %r8 # 8-byte Folded Reload movq %r8, 48(%rbx) movq %r9, 56(%rbx) addq $776, %rsp # imm = 0x308 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end116: .size mcl_fp_montRed8L, .Lfunc_end116-mcl_fp_montRed8L .globl mcl_fp_addPre8L .align 16, 0x90 .type mcl_fp_addPre8L,@function mcl_fp_addPre8L: # @mcl_fp_addPre8L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 56(%rdx), %r8 movq 56(%rsi), %r15 movq 48(%rdx), %r9 movq 48(%rsi), %r12 movq 40(%rdx), %r10 movq 32(%rdx), %r11 movq 24(%rdx), %r14 movq 16(%rdx), %rbx movq (%rdx), %rcx movq 8(%rdx), %rdx addq (%rsi), %rcx adcq 8(%rsi), %rdx adcq 16(%rsi), %rbx movq 40(%rsi), %r13 movq 24(%rsi), %rax movq 32(%rsi), %rsi movq %rcx, (%rdi) movq %rdx, 8(%rdi) movq %rbx, 16(%rdi) adcq %r14, %rax movq %rax, 24(%rdi) adcq %r11, %rsi movq %rsi, 32(%rdi) adcq %r10, %r13 movq %r13, 40(%rdi) adcq %r9, %r12 movq %r12, 48(%rdi) adcq %r8, %r15 movq %r15, 56(%rdi) sbbq %rax, %rax andl $1, %eax popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end117: .size mcl_fp_addPre8L, .Lfunc_end117-mcl_fp_addPre8L .globl mcl_fp_subPre8L .align 16, 0x90 .type mcl_fp_subPre8L,@function mcl_fp_subPre8L: # @mcl_fp_subPre8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 56(%rdx), %r8 movq 56(%rsi), %r15 movq 48(%rdx), %r9 movq 40(%rdx), %r10 movq 24(%rdx), %r11 movq 32(%rdx), %r14 movq (%rsi), %rbx movq 8(%rsi), %r12 xorl %eax, %eax subq (%rdx), %rbx sbbq 8(%rdx), %r12 movq 16(%rsi), %rcx sbbq 16(%rdx), %rcx movq 48(%rsi), %r13 movq 40(%rsi), %rdx movq 32(%rsi), %rbp movq 24(%rsi), %rsi movq %rbx, (%rdi) movq %r12, 8(%rdi) movq %rcx, 16(%rdi) sbbq %r11, %rsi movq %rsi, 24(%rdi) sbbq %r14, %rbp movq %rbp, 32(%rdi) sbbq %r10, %rdx movq %rdx, 40(%rdi) sbbq %r9, %r13 movq %r13, 48(%rdi) sbbq %r8, %r15 movq %r15, 56(%rdi) sbbq $0, %rax andl $1, %eax popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end118: .size mcl_fp_subPre8L, .Lfunc_end118-mcl_fp_subPre8L .globl mcl_fp_shr1_8L .align 16, 0x90 .type mcl_fp_shr1_8L,@function mcl_fp_shr1_8L: # @mcl_fp_shr1_8L # BB#0: movq 56(%rsi), %r8 movq 48(%rsi), %r9 movq 40(%rsi), %r10 movq 32(%rsi), %r11 movq 24(%rsi), %rcx movq 16(%rsi), %rdx movq (%rsi), %rax movq 8(%rsi), %rsi shrdq $1, %rsi, %rax movq %rax, (%rdi) shrdq $1, %rdx, %rsi movq %rsi, 8(%rdi) shrdq $1, %rcx, %rdx movq %rdx, 16(%rdi) shrdq $1, %r11, %rcx movq %rcx, 24(%rdi) shrdq $1, %r10, %r11 movq %r11, 32(%rdi) shrdq $1, %r9, %r10 movq %r10, 40(%rdi) shrdq $1, %r8, %r9 movq %r9, 48(%rdi) shrq %r8 movq %r8, 56(%rdi) retq .Lfunc_end119: .size mcl_fp_shr1_8L, .Lfunc_end119-mcl_fp_shr1_8L .globl mcl_fp_add8L .align 16, 0x90 .type mcl_fp_add8L,@function mcl_fp_add8L: # @mcl_fp_add8L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 56(%rdx), %r15 movq 56(%rsi), %r8 movq 48(%rdx), %r12 movq 48(%rsi), %r9 movq 40(%rsi), %r13 movq 24(%rsi), %r11 movq 32(%rsi), %r10 movq (%rdx), %r14 movq 8(%rdx), %rbx addq (%rsi), %r14 adcq 8(%rsi), %rbx movq 16(%rdx), %rax adcq 16(%rsi), %rax adcq 24(%rdx), %r11 movq 40(%rdx), %rsi adcq 32(%rdx), %r10 movq %r14, (%rdi) movq %rbx, 8(%rdi) movq %rax, 16(%rdi) movq %r11, 24(%rdi) movq %r10, 32(%rdi) adcq %r13, %rsi movq %rsi, 40(%rdi) adcq %r12, %r9 movq %r9, 48(%rdi) adcq %r15, %r8 movq %r8, 56(%rdi) sbbq %rdx, %rdx andl $1, %edx subq (%rcx), %r14 sbbq 8(%rcx), %rbx sbbq 16(%rcx), %rax sbbq 24(%rcx), %r11 sbbq 32(%rcx), %r10 sbbq 40(%rcx), %rsi sbbq 48(%rcx), %r9 sbbq 56(%rcx), %r8 sbbq $0, %rdx testb $1, %dl jne .LBB120_2 # BB#1: # %nocarry movq %r14, (%rdi) movq %rbx, 8(%rdi) movq %rax, 16(%rdi) movq %r11, 24(%rdi) movq %r10, 32(%rdi) movq %rsi, 40(%rdi) movq %r9, 48(%rdi) movq %r8, 56(%rdi) .LBB120_2: # %carry popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end120: .size mcl_fp_add8L, .Lfunc_end120-mcl_fp_add8L .globl mcl_fp_addNF8L .align 16, 0x90 .type mcl_fp_addNF8L,@function mcl_fp_addNF8L: # @mcl_fp_addNF8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 56(%rdx), %r8 movq 48(%rdx), %rbp movq 40(%rdx), %rbx movq 32(%rdx), %rax movq 24(%rdx), %r11 movq 16(%rdx), %r15 movq (%rdx), %r13 movq 8(%rdx), %r12 addq (%rsi), %r13 adcq 8(%rsi), %r12 adcq 16(%rsi), %r15 adcq 24(%rsi), %r11 adcq 32(%rsi), %rax movq %rax, -24(%rsp) # 8-byte Spill movq %rax, %r10 adcq 40(%rsi), %rbx movq %rbx, -16(%rsp) # 8-byte Spill movq %rbx, %r9 adcq 48(%rsi), %rbp movq %rbp, -8(%rsp) # 8-byte Spill movq %rbp, %rax adcq 56(%rsi), %r8 movq %r13, %rsi subq (%rcx), %rsi movq %r12, %rdx sbbq 8(%rcx), %rdx movq %r15, %rbx sbbq 16(%rcx), %rbx movq %r11, %r14 sbbq 24(%rcx), %r14 movq %r10, %rbp sbbq 32(%rcx), %rbp movq %r9, %r10 sbbq 40(%rcx), %r10 movq %rax, %r9 sbbq 48(%rcx), %r9 movq %r8, %rax sbbq 56(%rcx), %rax testq %rax, %rax cmovsq %r13, %rsi movq %rsi, (%rdi) cmovsq %r12, %rdx movq %rdx, 8(%rdi) cmovsq %r15, %rbx movq %rbx, 16(%rdi) cmovsq %r11, %r14 movq %r14, 24(%rdi) cmovsq -24(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 32(%rdi) cmovsq -16(%rsp), %r10 # 8-byte Folded Reload movq %r10, 40(%rdi) cmovsq -8(%rsp), %r9 # 8-byte Folded Reload movq %r9, 48(%rdi) cmovsq %r8, %rax movq %rax, 56(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end121: .size mcl_fp_addNF8L, .Lfunc_end121-mcl_fp_addNF8L .globl mcl_fp_sub8L .align 16, 0x90 .type mcl_fp_sub8L,@function mcl_fp_sub8L: # @mcl_fp_sub8L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 56(%rdx), %r12 movq 56(%rsi), %r8 movq 48(%rdx), %r13 movq (%rsi), %rax movq 8(%rsi), %r10 xorl %ebx, %ebx subq (%rdx), %rax sbbq 8(%rdx), %r10 movq 16(%rsi), %r11 sbbq 16(%rdx), %r11 movq 24(%rsi), %r15 sbbq 24(%rdx), %r15 movq 32(%rsi), %r14 sbbq 32(%rdx), %r14 movq 48(%rsi), %r9 movq 40(%rsi), %rsi sbbq 40(%rdx), %rsi movq %rax, (%rdi) movq %r10, 8(%rdi) movq %r11, 16(%rdi) movq %r15, 24(%rdi) movq %r14, 32(%rdi) movq %rsi, 40(%rdi) sbbq %r13, %r9 movq %r9, 48(%rdi) sbbq %r12, %r8 movq %r8, 56(%rdi) sbbq $0, %rbx testb $1, %bl je .LBB122_2 # BB#1: # %carry addq (%rcx), %rax movq %rax, (%rdi) movq 8(%rcx), %rax adcq %r10, %rax movq %rax, 8(%rdi) movq 16(%rcx), %rax adcq %r11, %rax movq %rax, 16(%rdi) movq 24(%rcx), %rax adcq %r15, %rax movq %rax, 24(%rdi) movq 32(%rcx), %rax adcq %r14, %rax movq %rax, 32(%rdi) movq 40(%rcx), %rax adcq %rsi, %rax movq %rax, 40(%rdi) movq 48(%rcx), %rax adcq %r9, %rax movq %rax, 48(%rdi) movq 56(%rcx), %rax adcq %r8, %rax movq %rax, 56(%rdi) .LBB122_2: # %nocarry popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end122: .size mcl_fp_sub8L, .Lfunc_end122-mcl_fp_sub8L .globl mcl_fp_subNF8L .align 16, 0x90 .type mcl_fp_subNF8L,@function mcl_fp_subNF8L: # @mcl_fp_subNF8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r8 movq %rdi, %r9 movq 56(%rsi), %r14 movq 48(%rsi), %rax movq 40(%rsi), %rcx movq 32(%rsi), %rdi movq 24(%rsi), %r11 movq 16(%rsi), %r15 movq (%rsi), %r13 movq 8(%rsi), %r12 subq (%rdx), %r13 sbbq 8(%rdx), %r12 sbbq 16(%rdx), %r15 sbbq 24(%rdx), %r11 sbbq 32(%rdx), %rdi movq %rdi, -24(%rsp) # 8-byte Spill sbbq 40(%rdx), %rcx movq %rcx, -16(%rsp) # 8-byte Spill sbbq 48(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill sbbq 56(%rdx), %r14 movq %r14, %rsi sarq $63, %rsi movq 56(%r8), %r10 andq %rsi, %r10 movq 48(%r8), %rbx andq %rsi, %rbx movq 40(%r8), %rdi andq %rsi, %rdi movq 32(%r8), %rbp andq %rsi, %rbp movq 24(%r8), %rdx andq %rsi, %rdx movq 16(%r8), %rcx andq %rsi, %rcx movq 8(%r8), %rax andq %rsi, %rax andq (%r8), %rsi addq %r13, %rsi adcq %r12, %rax movq %rsi, (%r9) adcq %r15, %rcx movq %rax, 8(%r9) movq %rcx, 16(%r9) adcq %r11, %rdx movq %rdx, 24(%r9) adcq -24(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 32(%r9) adcq -16(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 40(%r9) adcq -8(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 48(%r9) adcq %r14, %r10 movq %r10, 56(%r9) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end123: .size mcl_fp_subNF8L, .Lfunc_end123-mcl_fp_subNF8L .globl mcl_fpDbl_add8L .align 16, 0x90 .type mcl_fpDbl_add8L,@function mcl_fpDbl_add8L: # @mcl_fpDbl_add8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r8 movq 120(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 112(%rdx), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 104(%rdx), %rax movq %rax, -32(%rsp) # 8-byte Spill movq 96(%rdx), %r14 movq 24(%rsi), %r15 movq 32(%rsi), %r11 movq 16(%rdx), %r12 movq (%rdx), %rbx movq 8(%rdx), %rax addq (%rsi), %rbx adcq 8(%rsi), %rax adcq 16(%rsi), %r12 adcq 24(%rdx), %r15 adcq 32(%rdx), %r11 movq 88(%rdx), %rbp movq 80(%rdx), %r13 movq %rbx, (%rdi) movq 72(%rdx), %r10 movq %rax, 8(%rdi) movq 64(%rdx), %r9 movq %r12, 16(%rdi) movq 40(%rdx), %r12 movq %r15, 24(%rdi) movq 40(%rsi), %rbx adcq %r12, %rbx movq 56(%rdx), %r15 movq 48(%rdx), %r12 movq %r11, 32(%rdi) movq 48(%rsi), %rdx adcq %r12, %rdx movq 120(%rsi), %r12 movq %rbx, 40(%rdi) movq 56(%rsi), %rax adcq %r15, %rax movq 112(%rsi), %rcx movq %rdx, 48(%rdi) movq 64(%rsi), %rbx adcq %r9, %rbx movq 104(%rsi), %rdx movq %rax, 56(%rdi) movq 72(%rsi), %r9 adcq %r10, %r9 movq 80(%rsi), %r11 adcq %r13, %r11 movq 96(%rsi), %rax movq 88(%rsi), %r15 adcq %rbp, %r15 adcq %r14, %rax movq %rax, -16(%rsp) # 8-byte Spill movq %rdx, %rax adcq -32(%rsp), %rax # 8-byte Folded Reload movq %rax, -32(%rsp) # 8-byte Spill adcq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -24(%rsp) # 8-byte Spill adcq -8(%rsp), %r12 # 8-byte Folded Reload movq %r12, -8(%rsp) # 8-byte Spill sbbq %rbp, %rbp andl $1, %ebp movq %rbx, %rsi subq (%r8), %rsi movq %r9, %rdx sbbq 8(%r8), %rdx movq %r11, %r10 sbbq 16(%r8), %r10 movq %r15, %r14 sbbq 24(%r8), %r14 movq -16(%rsp), %r13 # 8-byte Reload sbbq 32(%r8), %r13 movq %rax, %r12 sbbq 40(%r8), %r12 movq %rcx, %rax sbbq 48(%r8), %rax movq -8(%rsp), %rcx # 8-byte Reload sbbq 56(%r8), %rcx sbbq $0, %rbp andl $1, %ebp cmovneq %rbx, %rsi movq %rsi, 64(%rdi) testb %bpl, %bpl cmovneq %r9, %rdx movq %rdx, 72(%rdi) cmovneq %r11, %r10 movq %r10, 80(%rdi) cmovneq %r15, %r14 movq %r14, 88(%rdi) cmovneq -16(%rsp), %r13 # 8-byte Folded Reload movq %r13, 96(%rdi) cmovneq -32(%rsp), %r12 # 8-byte Folded Reload movq %r12, 104(%rdi) cmovneq -24(%rsp), %rax # 8-byte Folded Reload movq %rax, 112(%rdi) cmovneq -8(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 120(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end124: .size mcl_fpDbl_add8L, .Lfunc_end124-mcl_fpDbl_add8L .globl mcl_fpDbl_sub8L .align 16, 0x90 .type mcl_fpDbl_sub8L,@function mcl_fpDbl_sub8L: # @mcl_fpDbl_sub8L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r15 movq 120(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 112(%rdx), %rax movq %rax, -16(%rsp) # 8-byte Spill movq 104(%rdx), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 16(%rsi), %r9 movq (%rsi), %r12 movq 8(%rsi), %r14 xorl %r8d, %r8d subq (%rdx), %r12 sbbq 8(%rdx), %r14 sbbq 16(%rdx), %r9 movq 24(%rsi), %rbx sbbq 24(%rdx), %rbx movq 32(%rsi), %r13 sbbq 32(%rdx), %r13 movq 96(%rdx), %rbp movq 88(%rdx), %r11 movq %r12, (%rdi) movq 80(%rdx), %r12 movq %r14, 8(%rdi) movq 72(%rdx), %r10 movq %r9, 16(%rdi) movq 40(%rdx), %r9 movq %rbx, 24(%rdi) movq 40(%rsi), %rbx sbbq %r9, %rbx movq 48(%rdx), %r9 movq %r13, 32(%rdi) movq 48(%rsi), %r14 sbbq %r9, %r14 movq 64(%rdx), %r13 movq 56(%rdx), %r9 movq %rbx, 40(%rdi) movq 56(%rsi), %rdx sbbq %r9, %rdx movq 120(%rsi), %rcx movq %r14, 48(%rdi) movq 64(%rsi), %rbx sbbq %r13, %rbx movq 112(%rsi), %rax movq %rdx, 56(%rdi) movq 72(%rsi), %r9 sbbq %r10, %r9 movq 80(%rsi), %r13 sbbq %r12, %r13 movq 88(%rsi), %r12 sbbq %r11, %r12 movq 104(%rsi), %rdx movq 96(%rsi), %r14 sbbq %rbp, %r14 sbbq -24(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -24(%rsp) # 8-byte Spill sbbq -16(%rsp), %rax # 8-byte Folded Reload movq %rax, -16(%rsp) # 8-byte Spill sbbq -8(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -8(%rsp) # 8-byte Spill movl $0, %ebp sbbq $0, %rbp andl $1, %ebp movq (%r15), %r11 cmoveq %r8, %r11 testb %bpl, %bpl movq 16(%r15), %rbp cmoveq %r8, %rbp movq 8(%r15), %rsi cmoveq %r8, %rsi movq 56(%r15), %r10 cmoveq %r8, %r10 movq 48(%r15), %rdx cmoveq %r8, %rdx movq 40(%r15), %rcx cmoveq %r8, %rcx movq 32(%r15), %rax cmoveq %r8, %rax cmovneq 24(%r15), %r8 addq %rbx, %r11 adcq %r9, %rsi movq %r11, 64(%rdi) adcq %r13, %rbp movq %rsi, 72(%rdi) movq %rbp, 80(%rdi) adcq %r12, %r8 movq %r8, 88(%rdi) adcq %r14, %rax movq %rax, 96(%rdi) adcq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 104(%rdi) adcq -16(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 112(%rdi) adcq -8(%rsp), %r10 # 8-byte Folded Reload movq %r10, 120(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end125: .size mcl_fpDbl_sub8L, .Lfunc_end125-mcl_fpDbl_sub8L .align 16, 0x90 .type .LmulPv576x64,@function .LmulPv576x64: # @mulPv576x64 # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdx, %rbx movq %rbx, %rax mulq (%rsi) movq %rdx, -32(%rsp) # 8-byte Spill movq %rax, (%rdi) movq %rbx, %rax mulq 64(%rsi) movq %rdx, %r10 movq %rax, -8(%rsp) # 8-byte Spill movq %rbx, %rax mulq 56(%rsi) movq %rdx, %r14 movq %rax, -16(%rsp) # 8-byte Spill movq %rbx, %rax mulq 48(%rsi) movq %rdx, %r12 movq %rax, -24(%rsp) # 8-byte Spill movq %rbx, %rax mulq 40(%rsi) movq %rdx, %rcx movq %rax, -40(%rsp) # 8-byte Spill movq %rbx, %rax mulq 32(%rsi) movq %rdx, %rbp movq %rax, %r8 movq %rbx, %rax mulq 24(%rsi) movq %rdx, %r9 movq %rax, %r11 movq %rbx, %rax mulq 16(%rsi) movq %rdx, %r15 movq %rax, %r13 movq %rbx, %rax mulq 8(%rsi) addq -32(%rsp), %rax # 8-byte Folded Reload movq %rax, 8(%rdi) adcq %r13, %rdx movq %rdx, 16(%rdi) adcq %r11, %r15 movq %r15, 24(%rdi) adcq %r8, %r9 movq %r9, 32(%rdi) adcq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 40(%rdi) adcq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 48(%rdi) adcq -16(%rsp), %r12 # 8-byte Folded Reload movq %r12, 56(%rdi) adcq -8(%rsp), %r14 # 8-byte Folded Reload movq %r14, 64(%rdi) adcq $0, %r10 movq %r10, 72(%rdi) movq %rdi, %rax popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end126: .size .LmulPv576x64, .Lfunc_end126-.LmulPv576x64 .globl mcl_fp_mulUnitPre9L .align 16, 0x90 .type mcl_fp_mulUnitPre9L,@function mcl_fp_mulUnitPre9L: # @mcl_fp_mulUnitPre9L # BB#0: pushq %r14 pushq %rbx subq $88, %rsp movq %rdi, %rbx leaq 8(%rsp), %rdi callq .LmulPv576x64 movq 80(%rsp), %r8 movq 72(%rsp), %r9 movq 64(%rsp), %r10 movq 56(%rsp), %r11 movq 48(%rsp), %r14 movq 40(%rsp), %rax movq 32(%rsp), %rcx movq 24(%rsp), %rdx movq 8(%rsp), %rsi movq 16(%rsp), %rdi movq %rsi, (%rbx) movq %rdi, 8(%rbx) movq %rdx, 16(%rbx) movq %rcx, 24(%rbx) movq %rax, 32(%rbx) movq %r14, 40(%rbx) movq %r11, 48(%rbx) movq %r10, 56(%rbx) movq %r9, 64(%rbx) movq %r8, 72(%rbx) addq $88, %rsp popq %rbx popq %r14 retq .Lfunc_end127: .size mcl_fp_mulUnitPre9L, .Lfunc_end127-mcl_fp_mulUnitPre9L .globl mcl_fpDbl_mulPre9L .align 16, 0x90 .type mcl_fpDbl_mulPre9L,@function mcl_fpDbl_mulPre9L: # @mcl_fpDbl_mulPre9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $808, %rsp # imm = 0x328 movq %rdx, %rax movq %rax, 64(%rsp) # 8-byte Spill movq %rsi, 72(%rsp) # 8-byte Spill movq %rdi, %r12 movq %r12, 80(%rsp) # 8-byte Spill movq (%rax), %rdx movq %rax, %rbx leaq 728(%rsp), %rdi movq %rsi, %rbp callq .LmulPv576x64 movq 800(%rsp), %r13 movq 792(%rsp), %rax movq %rax, 56(%rsp) # 8-byte Spill movq 784(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 776(%rsp), %rax movq %rax, 40(%rsp) # 8-byte Spill movq 768(%rsp), %rax movq %rax, 32(%rsp) # 8-byte Spill movq 760(%rsp), %rax movq %rax, 24(%rsp) # 8-byte Spill movq 752(%rsp), %rax movq %rax, 16(%rsp) # 8-byte Spill movq 744(%rsp), %rax movq %rax, 8(%rsp) # 8-byte Spill movq 728(%rsp), %rax movq 736(%rsp), %r14 movq %rax, (%r12) movq 8(%rbx), %rdx leaq 648(%rsp), %rdi movq %rbp, %rsi callq .LmulPv576x64 movq 720(%rsp), %r8 movq 712(%rsp), %rcx movq 704(%rsp), %rdx movq 696(%rsp), %rsi movq 688(%rsp), %rdi movq 680(%rsp), %rbp addq 648(%rsp), %r14 movq 672(%rsp), %rax movq 656(%rsp), %rbx movq 664(%rsp), %r15 movq %r14, 8(%r12) adcq 8(%rsp), %rbx # 8-byte Folded Reload adcq 16(%rsp), %r15 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, %r14 adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 24(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 32(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 40(%rsp) # 8-byte Spill adcq %r13, %rcx movq %rcx, 48(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 56(%rsp) # 8-byte Spill movq 64(%rsp), %r13 # 8-byte Reload movq 16(%r13), %rdx leaq 568(%rsp), %rdi movq 72(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 640(%rsp), %r8 movq 632(%rsp), %r9 movq 624(%rsp), %r10 movq 616(%rsp), %rdi movq 608(%rsp), %rbp movq 600(%rsp), %rcx addq 568(%rsp), %rbx movq 592(%rsp), %rdx movq 576(%rsp), %r12 movq 584(%rsp), %rsi movq 80(%rsp), %rax # 8-byte Reload movq %rbx, 16(%rax) adcq %r15, %r12 adcq %r14, %rsi movq %rsi, (%rsp) # 8-byte Spill adcq 16(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 8(%rsp) # 8-byte Spill adcq 24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %r10 # 8-byte Folded Reload movq %r10, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %r9 # 8-byte Folded Reload movq %r9, 48(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 56(%rsp) # 8-byte Spill movq 24(%r13), %rdx leaq 488(%rsp), %rdi movq 72(%rsp), %r15 # 8-byte Reload movq %r15, %rsi callq .LmulPv576x64 movq 560(%rsp), %r8 movq 552(%rsp), %rcx movq 544(%rsp), %rdx movq 536(%rsp), %rsi movq 528(%rsp), %rdi movq 520(%rsp), %rbp addq 488(%rsp), %r12 movq 512(%rsp), %rax movq 496(%rsp), %rbx movq 504(%rsp), %r13 movq 80(%rsp), %r14 # 8-byte Reload movq %r12, 24(%r14) adcq (%rsp), %rbx # 8-byte Folded Reload adcq 8(%rsp), %r13 # 8-byte Folded Reload adcq 16(%rsp), %rax # 8-byte Folded Reload movq %rax, 8(%rsp) # 8-byte Spill adcq 24(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 48(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 56(%rsp) # 8-byte Spill movq 64(%rsp), %r12 # 8-byte Reload movq 32(%r12), %rdx leaq 408(%rsp), %rdi movq %r15, %rsi callq .LmulPv576x64 movq 480(%rsp), %r8 movq 472(%rsp), %r9 movq 464(%rsp), %rdx movq 456(%rsp), %rsi movq 448(%rsp), %rdi movq 440(%rsp), %rbp addq 408(%rsp), %rbx movq 432(%rsp), %rax movq 416(%rsp), %r15 movq 424(%rsp), %rcx movq %rbx, 32(%r14) adcq %r13, %r15 adcq 8(%rsp), %rcx # 8-byte Folded Reload movq %rcx, (%rsp) # 8-byte Spill adcq 16(%rsp), %rax # 8-byte Folded Reload movq %rax, 8(%rsp) # 8-byte Spill adcq 24(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %r9 # 8-byte Folded Reload movq %r9, 48(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 56(%rsp) # 8-byte Spill movq %r12, %r14 movq 40(%r14), %rdx leaq 328(%rsp), %rdi movq 72(%rsp), %r13 # 8-byte Reload movq %r13, %rsi callq .LmulPv576x64 movq 400(%rsp), %r8 movq 392(%rsp), %r9 movq 384(%rsp), %rsi movq 376(%rsp), %rdi movq 368(%rsp), %rbx movq 360(%rsp), %rbp addq 328(%rsp), %r15 movq 352(%rsp), %rcx movq 336(%rsp), %r12 movq 344(%rsp), %rdx movq 80(%rsp), %rax # 8-byte Reload movq %r15, 40(%rax) adcq (%rsp), %r12 # 8-byte Folded Reload adcq 8(%rsp), %rdx # 8-byte Folded Reload movq %rdx, (%rsp) # 8-byte Spill adcq 16(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 8(%rsp) # 8-byte Spill adcq 24(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %r9 # 8-byte Folded Reload movq %r9, 48(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 56(%rsp) # 8-byte Spill movq 48(%r14), %rdx leaq 248(%rsp), %rdi movq %r13, %rsi movq %r13, %r15 callq .LmulPv576x64 movq 320(%rsp), %r8 movq 312(%rsp), %r9 movq 304(%rsp), %rsi movq 296(%rsp), %rdi movq 288(%rsp), %rbx movq 280(%rsp), %rbp addq 248(%rsp), %r12 movq 272(%rsp), %rcx movq 256(%rsp), %r13 movq 264(%rsp), %rdx movq 80(%rsp), %rax # 8-byte Reload movq %r12, 48(%rax) adcq (%rsp), %r13 # 8-byte Folded Reload adcq 8(%rsp), %rdx # 8-byte Folded Reload movq %rdx, (%rsp) # 8-byte Spill adcq 16(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 8(%rsp) # 8-byte Spill adcq 24(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %r9 # 8-byte Folded Reload movq %r9, 48(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 56(%rsp) # 8-byte Spill movq 56(%r14), %rdx leaq 168(%rsp), %rdi movq %r15, %rsi callq .LmulPv576x64 movq 240(%rsp), %rcx movq 232(%rsp), %rdx movq 224(%rsp), %rsi movq 216(%rsp), %rdi movq 208(%rsp), %rbx addq 168(%rsp), %r13 movq 200(%rsp), %r12 movq 192(%rsp), %rbp movq 176(%rsp), %r14 movq 184(%rsp), %r15 movq 80(%rsp), %rax # 8-byte Reload movq %r13, 56(%rax) adcq (%rsp), %r14 # 8-byte Folded Reload adcq 8(%rsp), %r15 # 8-byte Folded Reload adcq 16(%rsp), %rbp # 8-byte Folded Reload adcq 24(%rsp), %r12 # 8-byte Folded Reload adcq 32(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %r13 adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq $0, %rcx movq %rcx, 56(%rsp) # 8-byte Spill movq 64(%rsp), %rax # 8-byte Reload movq 64(%rax), %rdx leaq 88(%rsp), %rdi movq 72(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 88(%rsp), %r14 adcq 96(%rsp), %r15 movq 160(%rsp), %r8 adcq 104(%rsp), %rbp movq 152(%rsp), %r9 movq 144(%rsp), %rdx movq 136(%rsp), %rsi movq 128(%rsp), %rdi movq 120(%rsp), %rbx movq 112(%rsp), %rax movq 80(%rsp), %rcx # 8-byte Reload movq %r14, 64(%rcx) movq %r15, 72(%rcx) adcq %r12, %rax movq %rbp, 80(%rcx) movq %rax, 88(%rcx) adcq %r13, %rbx movq %rbx, 96(%rcx) adcq 32(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 104(%rcx) adcq 40(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 112(%rcx) adcq 48(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 120(%rcx) adcq 56(%rsp), %r9 # 8-byte Folded Reload movq %r9, 128(%rcx) adcq $0, %r8 movq %r8, 136(%rcx) addq $808, %rsp # imm = 0x328 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end128: .size mcl_fpDbl_mulPre9L, .Lfunc_end128-mcl_fpDbl_mulPre9L .globl mcl_fpDbl_sqrPre9L .align 16, 0x90 .type mcl_fpDbl_sqrPre9L,@function mcl_fpDbl_sqrPre9L: # @mcl_fpDbl_sqrPre9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $808, %rsp # imm = 0x328 movq %rsi, %r15 movq %r15, 80(%rsp) # 8-byte Spill movq %rdi, %r14 movq %r14, 72(%rsp) # 8-byte Spill movq (%r15), %rdx leaq 728(%rsp), %rdi callq .LmulPv576x64 movq 800(%rsp), %rax movq %rax, 64(%rsp) # 8-byte Spill movq 792(%rsp), %rax movq %rax, 56(%rsp) # 8-byte Spill movq 784(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 776(%rsp), %rax movq %rax, 40(%rsp) # 8-byte Spill movq 768(%rsp), %rax movq %rax, 32(%rsp) # 8-byte Spill movq 760(%rsp), %rax movq %rax, 24(%rsp) # 8-byte Spill movq 752(%rsp), %rax movq %rax, 16(%rsp) # 8-byte Spill movq 744(%rsp), %rax movq %rax, 8(%rsp) # 8-byte Spill movq 728(%rsp), %rax movq 736(%rsp), %r12 movq %rax, (%r14) movq 8(%r15), %rdx leaq 648(%rsp), %rdi movq %r15, %rsi callq .LmulPv576x64 movq 720(%rsp), %r8 movq 712(%rsp), %rcx movq 704(%rsp), %rdx movq 696(%rsp), %rsi movq 688(%rsp), %rdi movq 680(%rsp), %rbp addq 648(%rsp), %r12 movq 672(%rsp), %rax movq 656(%rsp), %rbx movq 664(%rsp), %r13 movq %r12, 8(%r14) adcq 8(%rsp), %rbx # 8-byte Folded Reload adcq 16(%rsp), %r13 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 16(%r15), %rdx leaq 568(%rsp), %rdi movq %r15, %rsi callq .LmulPv576x64 movq 640(%rsp), %r8 movq 632(%rsp), %rcx movq 624(%rsp), %rdx movq 616(%rsp), %rsi movq 608(%rsp), %rdi movq 600(%rsp), %rbp addq 568(%rsp), %rbx movq 592(%rsp), %rax movq 576(%rsp), %r14 movq 584(%rsp), %r12 movq 72(%rsp), %r15 # 8-byte Reload movq %rbx, 16(%r15) adcq %r13, %r14 adcq 16(%rsp), %r12 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 80(%rsp), %rsi # 8-byte Reload movq 24(%rsi), %rdx leaq 488(%rsp), %rdi callq .LmulPv576x64 movq 560(%rsp), %r8 movq 552(%rsp), %rcx movq 544(%rsp), %rdx movq 536(%rsp), %rsi movq 528(%rsp), %rdi movq 520(%rsp), %rbp addq 488(%rsp), %r14 movq 512(%rsp), %rax movq 496(%rsp), %rbx movq 504(%rsp), %r13 movq %r14, 24(%r15) adcq %r12, %rbx adcq 16(%rsp), %r13 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 80(%rsp), %rsi # 8-byte Reload movq 32(%rsi), %rdx leaq 408(%rsp), %rdi callq .LmulPv576x64 movq 480(%rsp), %r8 movq 472(%rsp), %rcx movq 464(%rsp), %rdx movq 456(%rsp), %rsi movq 448(%rsp), %rdi movq 440(%rsp), %rbp addq 408(%rsp), %rbx movq 432(%rsp), %rax movq 416(%rsp), %r14 movq 424(%rsp), %r12 movq %rbx, 32(%r15) adcq %r13, %r14 adcq 16(%rsp), %r12 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 80(%rsp), %rsi # 8-byte Reload movq 40(%rsi), %rdx leaq 328(%rsp), %rdi callq .LmulPv576x64 movq 400(%rsp), %r8 movq 392(%rsp), %rcx movq 384(%rsp), %rdx movq 376(%rsp), %rsi movq 368(%rsp), %rdi movq 360(%rsp), %rbp addq 328(%rsp), %r14 movq 352(%rsp), %rax movq 336(%rsp), %rbx movq 344(%rsp), %r13 movq %r14, 40(%r15) adcq %r12, %rbx adcq 16(%rsp), %r13 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 80(%rsp), %rsi # 8-byte Reload movq 48(%rsi), %rdx leaq 248(%rsp), %rdi callq .LmulPv576x64 movq 320(%rsp), %r8 movq 312(%rsp), %rcx movq 304(%rsp), %rdx movq 296(%rsp), %rsi movq 288(%rsp), %rdi movq 280(%rsp), %rbp addq 248(%rsp), %rbx movq 272(%rsp), %rax movq 256(%rsp), %r12 movq 264(%rsp), %r14 movq %rbx, 48(%r15) adcq %r13, %r12 adcq 16(%rsp), %r14 # 8-byte Folded Reload adcq 24(%rsp), %rax # 8-byte Folded Reload movq %rax, 16(%rsp) # 8-byte Spill adcq 32(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%rsp) # 8-byte Spill adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rsp) # 8-byte Spill adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 80(%rsp), %rsi # 8-byte Reload movq 56(%rsi), %rdx leaq 168(%rsp), %rdi callq .LmulPv576x64 movq 240(%rsp), %r8 movq 232(%rsp), %rdx movq 224(%rsp), %rsi movq 216(%rsp), %rdi movq 208(%rsp), %rbx movq 200(%rsp), %rcx addq 168(%rsp), %r12 movq 192(%rsp), %r15 movq 176(%rsp), %r13 movq 184(%rsp), %rbp movq 72(%rsp), %rax # 8-byte Reload movq %r12, 56(%rax) adcq %r14, %r13 adcq 16(%rsp), %rbp # 8-byte Folded Reload adcq 24(%rsp), %r15 # 8-byte Folded Reload adcq 32(%rsp), %rcx # 8-byte Folded Reload movq %rcx, %r12 adcq 40(%rsp), %rbx # 8-byte Folded Reload movq %rbx, %r14 adcq 48(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 40(%rsp) # 8-byte Spill adcq 56(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 48(%rsp) # 8-byte Spill adcq 64(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 56(%rsp) # 8-byte Spill adcq $0, %r8 movq %r8, 64(%rsp) # 8-byte Spill movq 80(%rsp), %rsi # 8-byte Reload movq 64(%rsi), %rdx leaq 88(%rsp), %rdi callq .LmulPv576x64 addq 88(%rsp), %r13 adcq 96(%rsp), %rbp movq 160(%rsp), %r8 adcq 104(%rsp), %r15 movq 152(%rsp), %r9 movq 144(%rsp), %rdx movq 136(%rsp), %rsi movq 128(%rsp), %rdi movq 120(%rsp), %rbx movq 112(%rsp), %rax movq 72(%rsp), %rcx # 8-byte Reload movq %r13, 64(%rcx) movq %rbp, 72(%rcx) adcq %r12, %rax movq %r15, 80(%rcx) movq %rax, 88(%rcx) adcq %r14, %rbx movq %rbx, 96(%rcx) adcq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 104(%rcx) adcq 48(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 112(%rcx) adcq 56(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 120(%rcx) adcq 64(%rsp), %r9 # 8-byte Folded Reload movq %r9, 128(%rcx) adcq $0, %r8 movq %r8, 136(%rcx) addq $808, %rsp # imm = 0x328 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end129: .size mcl_fpDbl_sqrPre9L, .Lfunc_end129-mcl_fpDbl_sqrPre9L .globl mcl_fp_mont9L .align 16, 0x90 .type mcl_fp_mont9L,@function mcl_fp_mont9L: # @mcl_fp_mont9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $1560, %rsp # imm = 0x618 movq %rcx, 56(%rsp) # 8-byte Spill movq %rdx, 32(%rsp) # 8-byte Spill movq %rsi, 24(%rsp) # 8-byte Spill movq %rdi, (%rsp) # 8-byte Spill movq -8(%rcx), %rbx movq %rbx, 16(%rsp) # 8-byte Spill movq (%rdx), %rdx leaq 1480(%rsp), %rdi callq .LmulPv576x64 movq 1480(%rsp), %r14 movq 1488(%rsp), %r15 movq %r14, %rdx imulq %rbx, %rdx movq 1552(%rsp), %rax movq %rax, 112(%rsp) # 8-byte Spill movq 1544(%rsp), %rax movq %rax, 104(%rsp) # 8-byte Spill movq 1536(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 1528(%rsp), %r12 movq 1520(%rsp), %r13 movq 1512(%rsp), %rbx movq 1504(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 1496(%rsp), %rbp leaq 1400(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 1400(%rsp), %r14 adcq 1408(%rsp), %r15 adcq 1416(%rsp), %rbp movq %rbp, 96(%rsp) # 8-byte Spill movq 80(%rsp), %rax # 8-byte Reload adcq 1424(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill adcq 1432(%rsp), %rbx movq %rbx, 40(%rsp) # 8-byte Spill adcq 1440(%rsp), %r13 movq %r13, 64(%rsp) # 8-byte Spill adcq 1448(%rsp), %r12 movq %r12, 48(%rsp) # 8-byte Spill movq 88(%rsp), %rbx # 8-byte Reload adcq 1456(%rsp), %rbx movq 104(%rsp), %r14 # 8-byte Reload adcq 1464(%rsp), %r14 movq 112(%rsp), %r13 # 8-byte Reload adcq 1472(%rsp), %r13 sbbq %rbp, %rbp movq 32(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdx leaq 1320(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %ebp addq 1320(%rsp), %r15 movq 96(%rsp), %rax # 8-byte Reload adcq 1328(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill movq 80(%rsp), %rax # 8-byte Reload adcq 1336(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 40(%rsp), %r12 # 8-byte Reload adcq 1344(%rsp), %r12 movq 64(%rsp), %rax # 8-byte Reload adcq 1352(%rsp), %rax movq %rax, 64(%rsp) # 8-byte Spill movq 48(%rsp), %rax # 8-byte Reload adcq 1360(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill adcq 1368(%rsp), %rbx adcq 1376(%rsp), %r14 movq %r14, 104(%rsp) # 8-byte Spill adcq 1384(%rsp), %r13 movq %r13, 112(%rsp) # 8-byte Spill adcq 1392(%rsp), %rbp sbbq %r14, %r14 movq %r15, %rdx imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 1240(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq %r14, %rax andl $1, %eax addq 1240(%rsp), %r15 movq 96(%rsp), %rcx # 8-byte Reload adcq 1248(%rsp), %rcx movq %rcx, 96(%rsp) # 8-byte Spill movq 80(%rsp), %r14 # 8-byte Reload adcq 1256(%rsp), %r14 adcq 1264(%rsp), %r12 movq %r12, 40(%rsp) # 8-byte Spill movq 64(%rsp), %r12 # 8-byte Reload adcq 1272(%rsp), %r12 movq 48(%rsp), %r13 # 8-byte Reload adcq 1280(%rsp), %r13 adcq 1288(%rsp), %rbx movq %rbx, 88(%rsp) # 8-byte Spill movq 104(%rsp), %r15 # 8-byte Reload adcq 1296(%rsp), %r15 movq 112(%rsp), %rbx # 8-byte Reload adcq 1304(%rsp), %rbx adcq 1312(%rsp), %rbp adcq $0, %rax movq %rax, 72(%rsp) # 8-byte Spill movq 32(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdx leaq 1160(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 96(%rsp), %rax # 8-byte Reload addq 1160(%rsp), %rax adcq 1168(%rsp), %r14 movq %r14, 80(%rsp) # 8-byte Spill movq 40(%rsp), %r14 # 8-byte Reload adcq 1176(%rsp), %r14 adcq 1184(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill movq %r13, %r12 adcq 1192(%rsp), %r12 movq 88(%rsp), %rcx # 8-byte Reload adcq 1200(%rsp), %rcx movq %rcx, 88(%rsp) # 8-byte Spill adcq 1208(%rsp), %r15 movq %r15, %r13 adcq 1216(%rsp), %rbx movq %rbx, 112(%rsp) # 8-byte Spill adcq 1224(%rsp), %rbp movq 72(%rsp), %rcx # 8-byte Reload adcq 1232(%rsp), %rcx movq %rcx, 72(%rsp) # 8-byte Spill sbbq %r15, %r15 movq %rax, %rdx movq %rax, %rbx imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 1080(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq %r15, %rax andl $1, %eax addq 1080(%rsp), %rbx movq 80(%rsp), %rcx # 8-byte Reload adcq 1088(%rsp), %rcx movq %rcx, 80(%rsp) # 8-byte Spill movq %r14, %r15 adcq 1096(%rsp), %r15 movq 64(%rsp), %r14 # 8-byte Reload adcq 1104(%rsp), %r14 movq %r12, %rbx adcq 1112(%rsp), %rbx movq 88(%rsp), %rcx # 8-byte Reload adcq 1120(%rsp), %rcx movq %rcx, 88(%rsp) # 8-byte Spill adcq 1128(%rsp), %r13 movq %r13, 104(%rsp) # 8-byte Spill movq 112(%rsp), %r13 # 8-byte Reload adcq 1136(%rsp), %r13 adcq 1144(%rsp), %rbp movq 72(%rsp), %r12 # 8-byte Reload adcq 1152(%rsp), %r12 adcq $0, %rax movq %rax, 96(%rsp) # 8-byte Spill movq 32(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdx leaq 1000(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 80(%rsp), %rax # 8-byte Reload addq 1000(%rsp), %rax adcq 1008(%rsp), %r15 movq %r15, 40(%rsp) # 8-byte Spill adcq 1016(%rsp), %r14 movq %r14, %r15 adcq 1024(%rsp), %rbx movq %rbx, 48(%rsp) # 8-byte Spill movq 88(%rsp), %r14 # 8-byte Reload adcq 1032(%rsp), %r14 movq 104(%rsp), %rcx # 8-byte Reload adcq 1040(%rsp), %rcx movq %rcx, 104(%rsp) # 8-byte Spill adcq 1048(%rsp), %r13 movq %r13, 112(%rsp) # 8-byte Spill adcq 1056(%rsp), %rbp adcq 1064(%rsp), %r12 movq 96(%rsp), %rcx # 8-byte Reload adcq 1072(%rsp), %rcx movq %rcx, 96(%rsp) # 8-byte Spill sbbq %rbx, %rbx movq %rax, %rdx movq %rax, %r13 imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 920(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %ebx movq %rbx, %rax addq 920(%rsp), %r13 movq 40(%rsp), %rcx # 8-byte Reload adcq 928(%rsp), %rcx movq %rcx, 40(%rsp) # 8-byte Spill adcq 936(%rsp), %r15 movq %r15, 64(%rsp) # 8-byte Spill movq 48(%rsp), %r15 # 8-byte Reload adcq 944(%rsp), %r15 movq %r14, %r13 adcq 952(%rsp), %r13 movq 104(%rsp), %r14 # 8-byte Reload adcq 960(%rsp), %r14 movq 112(%rsp), %rbx # 8-byte Reload adcq 968(%rsp), %rbx adcq 976(%rsp), %rbp adcq 984(%rsp), %r12 movq 96(%rsp), %rcx # 8-byte Reload adcq 992(%rsp), %rcx movq %rcx, 96(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, 80(%rsp) # 8-byte Spill movq 32(%rsp), %rax # 8-byte Reload movq 32(%rax), %rdx leaq 840(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 40(%rsp), %rax # 8-byte Reload addq 840(%rsp), %rax movq 64(%rsp), %rcx # 8-byte Reload adcq 848(%rsp), %rcx movq %rcx, 64(%rsp) # 8-byte Spill adcq 856(%rsp), %r15 adcq 864(%rsp), %r13 movq %r13, 88(%rsp) # 8-byte Spill adcq 872(%rsp), %r14 movq %r14, 104(%rsp) # 8-byte Spill adcq 880(%rsp), %rbx movq %rbx, 112(%rsp) # 8-byte Spill adcq 888(%rsp), %rbp adcq 896(%rsp), %r12 movq 96(%rsp), %r13 # 8-byte Reload adcq 904(%rsp), %r13 movq 80(%rsp), %rcx # 8-byte Reload adcq 912(%rsp), %rcx movq %rcx, 80(%rsp) # 8-byte Spill sbbq %rbx, %rbx movq %rax, %rdx movq %rax, %r14 imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 760(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %ebx movq %rbx, %rax addq 760(%rsp), %r14 movq 64(%rsp), %rcx # 8-byte Reload adcq 768(%rsp), %rcx movq %rcx, 64(%rsp) # 8-byte Spill adcq 776(%rsp), %r15 movq 88(%rsp), %r14 # 8-byte Reload adcq 784(%rsp), %r14 movq 104(%rsp), %rcx # 8-byte Reload adcq 792(%rsp), %rcx movq %rcx, 104(%rsp) # 8-byte Spill movq 112(%rsp), %rcx # 8-byte Reload adcq 800(%rsp), %rcx movq %rcx, 112(%rsp) # 8-byte Spill adcq 808(%rsp), %rbp movq %r12, %rbx adcq 816(%rsp), %rbx movq %r13, %r12 adcq 824(%rsp), %r12 movq 80(%rsp), %r13 # 8-byte Reload adcq 832(%rsp), %r13 adcq $0, %rax movq %rax, 40(%rsp) # 8-byte Spill movq 32(%rsp), %rax # 8-byte Reload movq 40(%rax), %rdx leaq 680(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 64(%rsp), %rax # 8-byte Reload addq 680(%rsp), %rax adcq 688(%rsp), %r15 movq %r15, 48(%rsp) # 8-byte Spill adcq 696(%rsp), %r14 movq %r14, 88(%rsp) # 8-byte Spill movq 104(%rsp), %rcx # 8-byte Reload adcq 704(%rsp), %rcx movq %rcx, 104(%rsp) # 8-byte Spill movq 112(%rsp), %r15 # 8-byte Reload adcq 712(%rsp), %r15 adcq 720(%rsp), %rbp adcq 728(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill adcq 736(%rsp), %r12 movq %r12, 96(%rsp) # 8-byte Spill adcq 744(%rsp), %r13 movq %r13, 80(%rsp) # 8-byte Spill movq 40(%rsp), %r13 # 8-byte Reload adcq 752(%rsp), %r13 sbbq %r14, %r14 movq %rax, %rdx movq %rax, %rbx imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 600(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %r14d addq 600(%rsp), %rbx movq 48(%rsp), %rax # 8-byte Reload adcq 608(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 88(%rsp), %rax # 8-byte Reload adcq 616(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 104(%rsp), %rbx # 8-byte Reload adcq 624(%rsp), %rbx adcq 632(%rsp), %r15 movq %r15, 112(%rsp) # 8-byte Spill adcq 640(%rsp), %rbp movq 72(%rsp), %r12 # 8-byte Reload adcq 648(%rsp), %r12 movq 96(%rsp), %rax # 8-byte Reload adcq 656(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill movq 80(%rsp), %r15 # 8-byte Reload adcq 664(%rsp), %r15 adcq 672(%rsp), %r13 adcq $0, %r14 movq %r14, 64(%rsp) # 8-byte Spill movq 32(%rsp), %rax # 8-byte Reload movq 48(%rax), %rdx leaq 520(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 48(%rsp), %rax # 8-byte Reload addq 520(%rsp), %rax movq 88(%rsp), %r14 # 8-byte Reload adcq 528(%rsp), %r14 adcq 536(%rsp), %rbx movq %rbx, 104(%rsp) # 8-byte Spill movq 112(%rsp), %rcx # 8-byte Reload adcq 544(%rsp), %rcx movq %rcx, 112(%rsp) # 8-byte Spill adcq 552(%rsp), %rbp adcq 560(%rsp), %r12 movq %r12, 72(%rsp) # 8-byte Spill movq 96(%rsp), %r12 # 8-byte Reload adcq 568(%rsp), %r12 adcq 576(%rsp), %r15 movq %r15, 80(%rsp) # 8-byte Spill adcq 584(%rsp), %r13 movq %r13, 40(%rsp) # 8-byte Spill movq 64(%rsp), %r15 # 8-byte Reload adcq 592(%rsp), %r15 sbbq %rbx, %rbx movq %rax, %rdx movq %rax, %r13 imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 440(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %ebx movq %rbx, %rax addq 440(%rsp), %r13 adcq 448(%rsp), %r14 movq %r14, 88(%rsp) # 8-byte Spill movq 104(%rsp), %r14 # 8-byte Reload adcq 456(%rsp), %r14 movq 112(%rsp), %rbx # 8-byte Reload adcq 464(%rsp), %rbx adcq 472(%rsp), %rbp movq %rbp, 8(%rsp) # 8-byte Spill movq 72(%rsp), %rcx # 8-byte Reload adcq 480(%rsp), %rcx movq %rcx, 72(%rsp) # 8-byte Spill adcq 488(%rsp), %r12 movq %r12, 96(%rsp) # 8-byte Spill movq 80(%rsp), %rbp # 8-byte Reload adcq 496(%rsp), %rbp movq 40(%rsp), %r12 # 8-byte Reload adcq 504(%rsp), %r12 adcq 512(%rsp), %r15 movq %r15, %r13 adcq $0, %rax movq %rax, 48(%rsp) # 8-byte Spill movq 32(%rsp), %rax # 8-byte Reload movq 56(%rax), %rdx leaq 360(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 88(%rsp), %rax # 8-byte Reload addq 360(%rsp), %rax adcq 368(%rsp), %r14 adcq 376(%rsp), %rbx movq %rbx, 112(%rsp) # 8-byte Spill movq 8(%rsp), %rcx # 8-byte Reload adcq 384(%rsp), %rcx movq %rcx, 8(%rsp) # 8-byte Spill movq 72(%rsp), %rbx # 8-byte Reload adcq 392(%rsp), %rbx movq 96(%rsp), %r15 # 8-byte Reload adcq 400(%rsp), %r15 adcq 408(%rsp), %rbp movq %rbp, 80(%rsp) # 8-byte Spill adcq 416(%rsp), %r12 movq %r12, %rbp adcq 424(%rsp), %r13 movq %r13, 64(%rsp) # 8-byte Spill movq 48(%rsp), %rcx # 8-byte Reload adcq 432(%rsp), %rcx movq %rcx, 48(%rsp) # 8-byte Spill sbbq %r13, %r13 movq %rax, %rdx movq %rax, %r12 imulq 16(%rsp), %rdx # 8-byte Folded Reload leaq 280(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %r13d addq 280(%rsp), %r12 adcq 288(%rsp), %r14 movq %r14, 104(%rsp) # 8-byte Spill movq 112(%rsp), %rax # 8-byte Reload adcq 296(%rsp), %rax movq %rax, 112(%rsp) # 8-byte Spill movq 8(%rsp), %r14 # 8-byte Reload adcq 304(%rsp), %r14 adcq 312(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill adcq 320(%rsp), %r15 movq %r15, 96(%rsp) # 8-byte Spill movq 80(%rsp), %rbx # 8-byte Reload adcq 328(%rsp), %rbx adcq 336(%rsp), %rbp movq %rbp, 40(%rsp) # 8-byte Spill movq 64(%rsp), %r12 # 8-byte Reload adcq 344(%rsp), %r12 movq 48(%rsp), %rbp # 8-byte Reload adcq 352(%rsp), %rbp adcq $0, %r13 movq 32(%rsp), %rax # 8-byte Reload movq 64(%rax), %rdx leaq 200(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 104(%rsp), %rax # 8-byte Reload addq 200(%rsp), %rax movq 112(%rsp), %r15 # 8-byte Reload adcq 208(%rsp), %r15 adcq 216(%rsp), %r14 movq %r14, 8(%rsp) # 8-byte Spill movq 72(%rsp), %r14 # 8-byte Reload adcq 224(%rsp), %r14 movq 96(%rsp), %rcx # 8-byte Reload adcq 232(%rsp), %rcx movq %rcx, 96(%rsp) # 8-byte Spill adcq 240(%rsp), %rbx movq %rbx, 80(%rsp) # 8-byte Spill movq 40(%rsp), %rcx # 8-byte Reload adcq 248(%rsp), %rcx movq %rcx, 40(%rsp) # 8-byte Spill adcq 256(%rsp), %r12 movq %r12, 64(%rsp) # 8-byte Spill adcq 264(%rsp), %rbp movq %rbp, 48(%rsp) # 8-byte Spill adcq 272(%rsp), %r13 sbbq %rbx, %rbx movq 16(%rsp), %rdx # 8-byte Reload imulq %rax, %rdx movq %rax, %r12 leaq 120(%rsp), %rdi movq 56(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %ebx addq 120(%rsp), %r12 adcq 128(%rsp), %r15 movq 8(%rsp), %rbp # 8-byte Reload adcq 136(%rsp), %rbp movq %r14, %rcx adcq 144(%rsp), %rcx movq %rcx, 72(%rsp) # 8-byte Spill movq 96(%rsp), %r8 # 8-byte Reload adcq 152(%rsp), %r8 movq %r8, 96(%rsp) # 8-byte Spill movq 80(%rsp), %r9 # 8-byte Reload adcq 160(%rsp), %r9 movq %r9, 80(%rsp) # 8-byte Spill movq 40(%rsp), %r10 # 8-byte Reload adcq 168(%rsp), %r10 movq %r10, 40(%rsp) # 8-byte Spill movq 64(%rsp), %rdi # 8-byte Reload adcq 176(%rsp), %rdi movq %rdi, 64(%rsp) # 8-byte Spill movq 48(%rsp), %r14 # 8-byte Reload adcq 184(%rsp), %r14 adcq 192(%rsp), %r13 adcq $0, %rbx movq %r15, %rsi movq %r15, %r12 movq 56(%rsp), %rdx # 8-byte Reload subq (%rdx), %rsi movq %rbp, %rax movq %rbp, %r15 sbbq 8(%rdx), %rax movq %rcx, %rbp sbbq 16(%rdx), %rbp movq %r8, %rcx sbbq 24(%rdx), %rcx movq %r9, %r8 sbbq 32(%rdx), %r8 movq %r10, %r11 sbbq 40(%rdx), %r11 movq %rdi, %r10 sbbq 48(%rdx), %r10 movq %r14, %rdi sbbq 56(%rdx), %rdi movq %r13, %r9 sbbq 64(%rdx), %r9 sbbq $0, %rbx andl $1, %ebx cmovneq %r13, %r9 testb %bl, %bl cmovneq %r12, %rsi movq (%rsp), %rbx # 8-byte Reload movq %rsi, (%rbx) cmovneq %r15, %rax movq %rax, 8(%rbx) cmovneq 72(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rbx) cmovneq 96(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 24(%rbx) cmovneq 80(%rsp), %r8 # 8-byte Folded Reload movq %r8, 32(%rbx) cmovneq 40(%rsp), %r11 # 8-byte Folded Reload movq %r11, 40(%rbx) cmovneq 64(%rsp), %r10 # 8-byte Folded Reload movq %r10, 48(%rbx) cmovneq %r14, %rdi movq %rdi, 56(%rbx) movq %r9, 64(%rbx) addq $1560, %rsp # imm = 0x618 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end130: .size mcl_fp_mont9L, .Lfunc_end130-mcl_fp_mont9L .globl mcl_fp_montNF9L .align 16, 0x90 .type mcl_fp_montNF9L,@function mcl_fp_montNF9L: # @mcl_fp_montNF9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $1560, %rsp # imm = 0x618 movq %rcx, 64(%rsp) # 8-byte Spill movq %rdx, 16(%rsp) # 8-byte Spill movq %rsi, 24(%rsp) # 8-byte Spill movq %rdi, (%rsp) # 8-byte Spill movq -8(%rcx), %rbx movq %rbx, 32(%rsp) # 8-byte Spill movq (%rdx), %rdx leaq 1480(%rsp), %rdi callq .LmulPv576x64 movq 1480(%rsp), %r12 movq 1488(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq %r12, %rdx imulq %rbx, %rdx movq 1552(%rsp), %rax movq %rax, 112(%rsp) # 8-byte Spill movq 1544(%rsp), %r13 movq 1536(%rsp), %rax movq %rax, 72(%rsp) # 8-byte Spill movq 1528(%rsp), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 1520(%rsp), %r14 movq 1512(%rsp), %r15 movq 1504(%rsp), %rbx movq 1496(%rsp), %rbp leaq 1400(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 1400(%rsp), %r12 movq 88(%rsp), %rax # 8-byte Reload adcq 1408(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill adcq 1416(%rsp), %rbp movq %rbp, 8(%rsp) # 8-byte Spill adcq 1424(%rsp), %rbx movq %rbx, 104(%rsp) # 8-byte Spill adcq 1432(%rsp), %r15 movq %r15, 56(%rsp) # 8-byte Spill adcq 1440(%rsp), %r14 movq %r14, 40(%rsp) # 8-byte Spill movq 48(%rsp), %rbx # 8-byte Reload adcq 1448(%rsp), %rbx movq 72(%rsp), %r12 # 8-byte Reload adcq 1456(%rsp), %r12 adcq 1464(%rsp), %r13 movq %r13, 96(%rsp) # 8-byte Spill movq 112(%rsp), %rbp # 8-byte Reload adcq 1472(%rsp), %rbp movq 16(%rsp), %rax # 8-byte Reload movq 8(%rax), %rdx leaq 1320(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 1392(%rsp), %rax movq 88(%rsp), %rcx # 8-byte Reload addq 1320(%rsp), %rcx movq 8(%rsp), %r15 # 8-byte Reload adcq 1328(%rsp), %r15 movq 104(%rsp), %r14 # 8-byte Reload adcq 1336(%rsp), %r14 movq 56(%rsp), %rdx # 8-byte Reload adcq 1344(%rsp), %rdx movq %rdx, 56(%rsp) # 8-byte Spill movq 40(%rsp), %r13 # 8-byte Reload adcq 1352(%rsp), %r13 adcq 1360(%rsp), %rbx movq %rbx, 48(%rsp) # 8-byte Spill adcq 1368(%rsp), %r12 movq %r12, 72(%rsp) # 8-byte Spill movq 96(%rsp), %rdx # 8-byte Reload adcq 1376(%rsp), %rdx movq %rdx, 96(%rsp) # 8-byte Spill adcq 1384(%rsp), %rbp movq %rbp, 112(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, %rbp movq %rcx, %rdx movq %rcx, %rbx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 1240(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 1240(%rsp), %rbx adcq 1248(%rsp), %r15 movq %r15, 8(%rsp) # 8-byte Spill adcq 1256(%rsp), %r14 movq %r14, 104(%rsp) # 8-byte Spill movq 56(%rsp), %r12 # 8-byte Reload adcq 1264(%rsp), %r12 adcq 1272(%rsp), %r13 movq %r13, %r14 movq 48(%rsp), %r13 # 8-byte Reload adcq 1280(%rsp), %r13 movq 72(%rsp), %rbx # 8-byte Reload adcq 1288(%rsp), %rbx movq 96(%rsp), %r15 # 8-byte Reload adcq 1296(%rsp), %r15 movq 112(%rsp), %rax # 8-byte Reload adcq 1304(%rsp), %rax movq %rax, 112(%rsp) # 8-byte Spill adcq 1312(%rsp), %rbp movq %rbp, 80(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 16(%rax), %rdx leaq 1160(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 1232(%rsp), %rax movq 8(%rsp), %rcx # 8-byte Reload addq 1160(%rsp), %rcx movq 104(%rsp), %rbp # 8-byte Reload adcq 1168(%rsp), %rbp adcq 1176(%rsp), %r12 movq %r12, 56(%rsp) # 8-byte Spill adcq 1184(%rsp), %r14 adcq 1192(%rsp), %r13 movq %r13, %r12 adcq 1200(%rsp), %rbx movq %rbx, 72(%rsp) # 8-byte Spill adcq 1208(%rsp), %r15 movq %r15, 96(%rsp) # 8-byte Spill movq 112(%rsp), %rbx # 8-byte Reload adcq 1216(%rsp), %rbx movq 80(%rsp), %rdx # 8-byte Reload adcq 1224(%rsp), %rdx movq %rdx, 80(%rsp) # 8-byte Spill movq %rax, %r15 adcq $0, %r15 movq %rcx, %rdx movq %rcx, %r13 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 1080(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 1080(%rsp), %r13 adcq 1088(%rsp), %rbp movq %rbp, 104(%rsp) # 8-byte Spill movq 56(%rsp), %r13 # 8-byte Reload adcq 1096(%rsp), %r13 adcq 1104(%rsp), %r14 adcq 1112(%rsp), %r12 movq %r12, 48(%rsp) # 8-byte Spill movq 72(%rsp), %r12 # 8-byte Reload adcq 1120(%rsp), %r12 movq 96(%rsp), %rbp # 8-byte Reload adcq 1128(%rsp), %rbp adcq 1136(%rsp), %rbx movq %rbx, 112(%rsp) # 8-byte Spill movq 80(%rsp), %rbx # 8-byte Reload adcq 1144(%rsp), %rbx adcq 1152(%rsp), %r15 movq 16(%rsp), %rax # 8-byte Reload movq 24(%rax), %rdx leaq 1000(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 1072(%rsp), %rax movq 104(%rsp), %rcx # 8-byte Reload addq 1000(%rsp), %rcx adcq 1008(%rsp), %r13 movq %r13, 56(%rsp) # 8-byte Spill adcq 1016(%rsp), %r14 movq %r14, 40(%rsp) # 8-byte Spill movq 48(%rsp), %r14 # 8-byte Reload adcq 1024(%rsp), %r14 adcq 1032(%rsp), %r12 adcq 1040(%rsp), %rbp movq %rbp, 96(%rsp) # 8-byte Spill movq 112(%rsp), %r13 # 8-byte Reload adcq 1048(%rsp), %r13 adcq 1056(%rsp), %rbx movq %rbx, 80(%rsp) # 8-byte Spill adcq 1064(%rsp), %r15 movq %r15, 88(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, 104(%rsp) # 8-byte Spill movq %rcx, %rdx movq %rcx, %rbx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 920(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 920(%rsp), %rbx movq 56(%rsp), %rax # 8-byte Reload adcq 928(%rsp), %rax movq %rax, 56(%rsp) # 8-byte Spill movq 40(%rsp), %rbp # 8-byte Reload adcq 936(%rsp), %rbp movq %r14, %rbx adcq 944(%rsp), %rbx adcq 952(%rsp), %r12 movq 96(%rsp), %rax # 8-byte Reload adcq 960(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill adcq 968(%rsp), %r13 movq %r13, %r15 movq 80(%rsp), %r13 # 8-byte Reload adcq 976(%rsp), %r13 movq 88(%rsp), %r14 # 8-byte Reload adcq 984(%rsp), %r14 movq 104(%rsp), %rax # 8-byte Reload adcq 992(%rsp), %rax movq %rax, 104(%rsp) # 8-byte Spill movq 16(%rsp), %rax # 8-byte Reload movq 32(%rax), %rdx leaq 840(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 912(%rsp), %rax movq 56(%rsp), %rcx # 8-byte Reload addq 840(%rsp), %rcx adcq 848(%rsp), %rbp movq %rbp, 40(%rsp) # 8-byte Spill adcq 856(%rsp), %rbx movq %rbx, 48(%rsp) # 8-byte Spill adcq 864(%rsp), %r12 movq 96(%rsp), %rbp # 8-byte Reload adcq 872(%rsp), %rbp adcq 880(%rsp), %r15 movq %r15, 112(%rsp) # 8-byte Spill adcq 888(%rsp), %r13 adcq 896(%rsp), %r14 movq %r14, 88(%rsp) # 8-byte Spill movq 104(%rsp), %rdx # 8-byte Reload adcq 904(%rsp), %rdx movq %rdx, 104(%rsp) # 8-byte Spill adcq $0, %rax movq %rax, %r14 movq %rcx, %rdx movq %rcx, %rbx imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 760(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 760(%rsp), %rbx movq 40(%rsp), %rax # 8-byte Reload adcq 768(%rsp), %rax movq %rax, 40(%rsp) # 8-byte Spill movq 48(%rsp), %r15 # 8-byte Reload adcq 776(%rsp), %r15 adcq 784(%rsp), %r12 movq %r12, 72(%rsp) # 8-byte Spill movq %rbp, %rbx adcq 792(%rsp), %rbx movq 112(%rsp), %rbp # 8-byte Reload adcq 800(%rsp), %rbp adcq 808(%rsp), %r13 movq 88(%rsp), %rax # 8-byte Reload adcq 816(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 104(%rsp), %r12 # 8-byte Reload adcq 824(%rsp), %r12 adcq 832(%rsp), %r14 movq 16(%rsp), %rax # 8-byte Reload movq 40(%rax), %rdx leaq 680(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 752(%rsp), %rcx movq 40(%rsp), %rax # 8-byte Reload addq 680(%rsp), %rax adcq 688(%rsp), %r15 movq %r15, 48(%rsp) # 8-byte Spill movq 72(%rsp), %rdx # 8-byte Reload adcq 696(%rsp), %rdx movq %rdx, 72(%rsp) # 8-byte Spill adcq 704(%rsp), %rbx movq %rbx, 96(%rsp) # 8-byte Spill adcq 712(%rsp), %rbp movq %rbp, 112(%rsp) # 8-byte Spill adcq 720(%rsp), %r13 movq %r13, %r15 movq 88(%rsp), %rbx # 8-byte Reload adcq 728(%rsp), %rbx adcq 736(%rsp), %r12 movq %r12, 104(%rsp) # 8-byte Spill adcq 744(%rsp), %r14 movq %r14, 40(%rsp) # 8-byte Spill adcq $0, %rcx movq %rcx, 56(%rsp) # 8-byte Spill movq %rax, %rdx movq %rax, %r13 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 600(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 600(%rsp), %r13 movq 48(%rsp), %r13 # 8-byte Reload adcq 608(%rsp), %r13 movq 72(%rsp), %r12 # 8-byte Reload adcq 616(%rsp), %r12 movq 96(%rsp), %rbp # 8-byte Reload adcq 624(%rsp), %rbp movq 112(%rsp), %rax # 8-byte Reload adcq 632(%rsp), %rax movq %rax, 112(%rsp) # 8-byte Spill adcq 640(%rsp), %r15 movq %r15, 80(%rsp) # 8-byte Spill adcq 648(%rsp), %rbx movq %rbx, 88(%rsp) # 8-byte Spill movq 104(%rsp), %r14 # 8-byte Reload adcq 656(%rsp), %r14 movq 40(%rsp), %rbx # 8-byte Reload adcq 664(%rsp), %rbx movq 56(%rsp), %r15 # 8-byte Reload adcq 672(%rsp), %r15 movq 16(%rsp), %rax # 8-byte Reload movq 48(%rax), %rdx leaq 520(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 592(%rsp), %rcx movq %r13, %rax addq 520(%rsp), %rax adcq 528(%rsp), %r12 movq %r12, 72(%rsp) # 8-byte Spill movq %rbp, %r12 adcq 536(%rsp), %r12 movq 112(%rsp), %rbp # 8-byte Reload adcq 544(%rsp), %rbp movq 80(%rsp), %rdx # 8-byte Reload adcq 552(%rsp), %rdx movq %rdx, 80(%rsp) # 8-byte Spill movq 88(%rsp), %rdx # 8-byte Reload adcq 560(%rsp), %rdx movq %rdx, 88(%rsp) # 8-byte Spill adcq 568(%rsp), %r14 movq %r14, 104(%rsp) # 8-byte Spill adcq 576(%rsp), %rbx movq %rbx, 40(%rsp) # 8-byte Spill adcq 584(%rsp), %r15 movq %r15, 56(%rsp) # 8-byte Spill adcq $0, %rcx movq %rcx, %r13 movq %rax, %rdx movq %rax, %r14 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 440(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 440(%rsp), %r14 movq 72(%rsp), %rax # 8-byte Reload adcq 448(%rsp), %rax movq %rax, 72(%rsp) # 8-byte Spill adcq 456(%rsp), %r12 adcq 464(%rsp), %rbp movq %rbp, 112(%rsp) # 8-byte Spill movq 80(%rsp), %r14 # 8-byte Reload adcq 472(%rsp), %r14 movq 88(%rsp), %r15 # 8-byte Reload adcq 480(%rsp), %r15 movq 104(%rsp), %rbp # 8-byte Reload adcq 488(%rsp), %rbp movq 40(%rsp), %rbx # 8-byte Reload adcq 496(%rsp), %rbx movq 56(%rsp), %rax # 8-byte Reload adcq 504(%rsp), %rax movq %rax, 56(%rsp) # 8-byte Spill adcq 512(%rsp), %r13 movq 16(%rsp), %rax # 8-byte Reload movq 56(%rax), %rdx leaq 360(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 432(%rsp), %rcx movq 72(%rsp), %rax # 8-byte Reload addq 360(%rsp), %rax adcq 368(%rsp), %r12 movq %r12, 96(%rsp) # 8-byte Spill movq 112(%rsp), %rdx # 8-byte Reload adcq 376(%rsp), %rdx movq %rdx, 112(%rsp) # 8-byte Spill adcq 384(%rsp), %r14 movq %r14, 80(%rsp) # 8-byte Spill adcq 392(%rsp), %r15 movq %r15, 88(%rsp) # 8-byte Spill adcq 400(%rsp), %rbp movq %rbp, 104(%rsp) # 8-byte Spill adcq 408(%rsp), %rbx movq %rbx, 40(%rsp) # 8-byte Spill movq 56(%rsp), %r14 # 8-byte Reload adcq 416(%rsp), %r14 adcq 424(%rsp), %r13 movq %r13, %r15 adcq $0, %rcx movq %rcx, 72(%rsp) # 8-byte Spill movq %rax, %rdx movq %rax, %r12 imulq 32(%rsp), %rdx # 8-byte Folded Reload leaq 280(%rsp), %rdi movq 64(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 280(%rsp), %r12 movq 96(%rsp), %rax # 8-byte Reload adcq 288(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill movq 112(%rsp), %rbp # 8-byte Reload adcq 296(%rsp), %rbp movq 80(%rsp), %rax # 8-byte Reload adcq 304(%rsp), %rax movq %rax, 80(%rsp) # 8-byte Spill movq 88(%rsp), %r13 # 8-byte Reload adcq 312(%rsp), %r13 movq 104(%rsp), %r12 # 8-byte Reload adcq 320(%rsp), %r12 movq 40(%rsp), %rbx # 8-byte Reload adcq 328(%rsp), %rbx adcq 336(%rsp), %r14 movq %r14, 56(%rsp) # 8-byte Spill adcq 344(%rsp), %r15 movq %r15, 48(%rsp) # 8-byte Spill movq 72(%rsp), %r14 # 8-byte Reload adcq 352(%rsp), %r14 movq 16(%rsp), %rax # 8-byte Reload movq 64(%rax), %rdx leaq 200(%rsp), %rdi movq 24(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 movq 272(%rsp), %rcx movq 96(%rsp), %rax # 8-byte Reload addq 200(%rsp), %rax adcq 208(%rsp), %rbp movq %rbp, 112(%rsp) # 8-byte Spill movq 80(%rsp), %rbp # 8-byte Reload adcq 216(%rsp), %rbp adcq 224(%rsp), %r13 movq %r13, 88(%rsp) # 8-byte Spill adcq 232(%rsp), %r12 movq %r12, 104(%rsp) # 8-byte Spill adcq 240(%rsp), %rbx movq %rbx, 40(%rsp) # 8-byte Spill movq 56(%rsp), %r15 # 8-byte Reload adcq 248(%rsp), %r15 movq 48(%rsp), %r12 # 8-byte Reload adcq 256(%rsp), %r12 adcq 264(%rsp), %r14 adcq $0, %rcx movq %rcx, 96(%rsp) # 8-byte Spill movq 32(%rsp), %rdx # 8-byte Reload imulq %rax, %rdx movq %rax, %rbx leaq 120(%rsp), %rdi movq 64(%rsp), %r13 # 8-byte Reload movq %r13, %rsi callq .LmulPv576x64 addq 120(%rsp), %rbx movq 112(%rsp), %rcx # 8-byte Reload adcq 128(%rsp), %rcx movq %rbp, %rdx adcq 136(%rsp), %rdx movq 88(%rsp), %rsi # 8-byte Reload adcq 144(%rsp), %rsi movq %rsi, 88(%rsp) # 8-byte Spill movq 104(%rsp), %rdi # 8-byte Reload adcq 152(%rsp), %rdi movq %rdi, 104(%rsp) # 8-byte Spill movq 40(%rsp), %rbx # 8-byte Reload adcq 160(%rsp), %rbx movq %rbx, 40(%rsp) # 8-byte Spill movq %r15, %r8 adcq 168(%rsp), %r8 movq %r8, 56(%rsp) # 8-byte Spill movq %r12, %r15 adcq 176(%rsp), %r15 adcq 184(%rsp), %r14 movq 96(%rsp), %r9 # 8-byte Reload adcq 192(%rsp), %r9 movq %rcx, %rax movq %rcx, %r11 movq %r13, %rbp subq (%rbp), %rax movq %rdx, %rcx movq %rdx, %r12 sbbq 8(%rbp), %rcx movq %rsi, %rdx sbbq 16(%rbp), %rdx movq %rdi, %rsi sbbq 24(%rbp), %rsi movq %rbx, %rdi sbbq 32(%rbp), %rdi movq %r8, %r10 sbbq 40(%rbp), %r10 movq %r15, %r13 sbbq 48(%rbp), %r13 movq %r14, %r8 sbbq 56(%rbp), %r8 movq %rbp, %rbx movq %r9, %rbp sbbq 64(%rbx), %rbp movq %rbp, %rbx sarq $63, %rbx cmovsq %r11, %rax movq (%rsp), %rbx # 8-byte Reload movq %rax, (%rbx) cmovsq %r12, %rcx movq %rcx, 8(%rbx) cmovsq 88(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 16(%rbx) cmovsq 104(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 24(%rbx) cmovsq 40(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 32(%rbx) cmovsq 56(%rsp), %r10 # 8-byte Folded Reload movq %r10, 40(%rbx) cmovsq %r15, %r13 movq %r13, 48(%rbx) cmovsq %r14, %r8 movq %r8, 56(%rbx) cmovsq %r9, %rbp movq %rbp, 64(%rbx) addq $1560, %rsp # imm = 0x618 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end131: .size mcl_fp_montNF9L, .Lfunc_end131-mcl_fp_montNF9L .globl mcl_fp_montRed9L .align 16, 0x90 .type mcl_fp_montRed9L,@function mcl_fp_montRed9L: # @mcl_fp_montRed9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx subq $936, %rsp # imm = 0x3A8 movq %rdx, %rax movq %rax, 128(%rsp) # 8-byte Spill movq %rdi, 80(%rsp) # 8-byte Spill movq -8(%rax), %rcx movq %rcx, 120(%rsp) # 8-byte Spill movq (%rsi), %r14 movq 8(%rsi), %rdx movq %rdx, 192(%rsp) # 8-byte Spill movq %r14, %rdx imulq %rcx, %rdx movq 136(%rsi), %rcx movq %rcx, 112(%rsp) # 8-byte Spill movq 128(%rsi), %rcx movq %rcx, 152(%rsp) # 8-byte Spill movq 120(%rsi), %rcx movq %rcx, 104(%rsp) # 8-byte Spill movq 112(%rsi), %rcx movq %rcx, 144(%rsp) # 8-byte Spill movq 104(%rsi), %rcx movq %rcx, 184(%rsp) # 8-byte Spill movq 96(%rsi), %rcx movq %rcx, 208(%rsp) # 8-byte Spill movq 88(%rsi), %rcx movq %rcx, 200(%rsp) # 8-byte Spill movq 80(%rsi), %rcx movq %rcx, 160(%rsp) # 8-byte Spill movq 72(%rsi), %r12 movq 64(%rsi), %rcx movq %rcx, 176(%rsp) # 8-byte Spill movq 56(%rsi), %rcx movq %rcx, 168(%rsp) # 8-byte Spill movq 48(%rsi), %rcx movq %rcx, 136(%rsp) # 8-byte Spill movq 40(%rsi), %rbp movq 32(%rsi), %rbx movq 24(%rsi), %r13 movq 16(%rsi), %r15 movq %rax, %rcx movq (%rcx), %rax movq %rax, 16(%rsp) # 8-byte Spill movq 64(%rcx), %rax movq %rax, 72(%rsp) # 8-byte Spill movq 56(%rcx), %rax movq %rax, 64(%rsp) # 8-byte Spill movq 48(%rcx), %rax movq %rax, 56(%rsp) # 8-byte Spill movq 40(%rcx), %rax movq %rax, 48(%rsp) # 8-byte Spill movq 32(%rcx), %rax movq %rax, 40(%rsp) # 8-byte Spill movq 24(%rcx), %rax movq %rax, 32(%rsp) # 8-byte Spill movq 16(%rcx), %rax movq %rax, 24(%rsp) # 8-byte Spill movq 8(%rcx), %rax movq %rax, 8(%rsp) # 8-byte Spill movq %rcx, %rsi leaq 856(%rsp), %rdi callq .LmulPv576x64 addq 856(%rsp), %r14 movq 192(%rsp), %rcx # 8-byte Reload adcq 864(%rsp), %rcx adcq 872(%rsp), %r15 adcq 880(%rsp), %r13 adcq 888(%rsp), %rbx movq %rbx, 88(%rsp) # 8-byte Spill adcq 896(%rsp), %rbp movq %rbp, 96(%rsp) # 8-byte Spill movq 136(%rsp), %rax # 8-byte Reload adcq 904(%rsp), %rax movq %rax, 136(%rsp) # 8-byte Spill movq 168(%rsp), %rax # 8-byte Reload adcq 912(%rsp), %rax movq %rax, 168(%rsp) # 8-byte Spill movq 176(%rsp), %rax # 8-byte Reload adcq 920(%rsp), %rax movq %rax, 176(%rsp) # 8-byte Spill adcq 928(%rsp), %r12 movq %r12, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rbp # 8-byte Reload adcq $0, %rbp adcq $0, 200(%rsp) # 8-byte Folded Spill adcq $0, 208(%rsp) # 8-byte Folded Spill adcq $0, 184(%rsp) # 8-byte Folded Spill adcq $0, 144(%rsp) # 8-byte Folded Spill adcq $0, 104(%rsp) # 8-byte Folded Spill adcq $0, 152(%rsp) # 8-byte Folded Spill movq 112(%rsp), %r14 # 8-byte Reload adcq $0, %r14 sbbq %r12, %r12 movq %rcx, %rdx movq %rcx, %rbx imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 776(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 andl $1, %r12d addq 776(%rsp), %rbx adcq 784(%rsp), %r15 adcq 792(%rsp), %r13 movq %r13, (%rsp) # 8-byte Spill movq 88(%rsp), %rax # 8-byte Reload adcq 800(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 96(%rsp), %rax # 8-byte Reload adcq 808(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill movq 136(%rsp), %rax # 8-byte Reload adcq 816(%rsp), %rax movq %rax, 136(%rsp) # 8-byte Spill movq 168(%rsp), %rax # 8-byte Reload adcq 824(%rsp), %rax movq %rax, 168(%rsp) # 8-byte Spill movq 176(%rsp), %rax # 8-byte Reload adcq 832(%rsp), %rax movq %rax, 176(%rsp) # 8-byte Spill movq 192(%rsp), %rax # 8-byte Reload adcq 840(%rsp), %rax movq %rax, 192(%rsp) # 8-byte Spill adcq 848(%rsp), %rbp movq %rbp, 160(%rsp) # 8-byte Spill movq 200(%rsp), %r13 # 8-byte Reload adcq $0, %r13 adcq $0, 208(%rsp) # 8-byte Folded Spill adcq $0, 184(%rsp) # 8-byte Folded Spill adcq $0, 144(%rsp) # 8-byte Folded Spill adcq $0, 104(%rsp) # 8-byte Folded Spill movq 152(%rsp), %rbx # 8-byte Reload adcq $0, %rbx adcq $0, %r14 movq %r14, 112(%rsp) # 8-byte Spill adcq $0, %r12 movq %r15, %rdx imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 696(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 696(%rsp), %r15 movq (%rsp), %rcx # 8-byte Reload adcq 704(%rsp), %rcx movq 88(%rsp), %rax # 8-byte Reload adcq 712(%rsp), %rax movq %rax, 88(%rsp) # 8-byte Spill movq 96(%rsp), %rax # 8-byte Reload adcq 720(%rsp), %rax movq %rax, 96(%rsp) # 8-byte Spill movq 136(%rsp), %rbp # 8-byte Reload adcq 728(%rsp), %rbp movq 168(%rsp), %r14 # 8-byte Reload adcq 736(%rsp), %r14 movq 176(%rsp), %r15 # 8-byte Reload adcq 744(%rsp), %r15 movq 192(%rsp), %rax # 8-byte Reload adcq 752(%rsp), %rax movq %rax, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rax # 8-byte Reload adcq 760(%rsp), %rax movq %rax, 160(%rsp) # 8-byte Spill adcq 768(%rsp), %r13 movq %r13, 200(%rsp) # 8-byte Spill adcq $0, 208(%rsp) # 8-byte Folded Spill movq 184(%rsp), %r13 # 8-byte Reload adcq $0, %r13 adcq $0, 144(%rsp) # 8-byte Folded Spill adcq $0, 104(%rsp) # 8-byte Folded Spill adcq $0, %rbx movq %rbx, 152(%rsp) # 8-byte Spill adcq $0, 112(%rsp) # 8-byte Folded Spill adcq $0, %r12 movq %rcx, %rbx movq %rbx, %rdx imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 616(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 616(%rsp), %rbx movq 88(%rsp), %rax # 8-byte Reload adcq 624(%rsp), %rax movq 96(%rsp), %rcx # 8-byte Reload adcq 632(%rsp), %rcx movq %rcx, 96(%rsp) # 8-byte Spill adcq 640(%rsp), %rbp movq %rbp, 136(%rsp) # 8-byte Spill adcq 648(%rsp), %r14 movq %r14, 168(%rsp) # 8-byte Spill adcq 656(%rsp), %r15 movq 192(%rsp), %r14 # 8-byte Reload adcq 664(%rsp), %r14 movq 160(%rsp), %rbp # 8-byte Reload adcq 672(%rsp), %rbp movq 200(%rsp), %rcx # 8-byte Reload adcq 680(%rsp), %rcx movq %rcx, 200(%rsp) # 8-byte Spill movq 208(%rsp), %rcx # 8-byte Reload adcq 688(%rsp), %rcx movq %rcx, 208(%rsp) # 8-byte Spill adcq $0, %r13 movq %r13, 184(%rsp) # 8-byte Spill adcq $0, 144(%rsp) # 8-byte Folded Spill adcq $0, 104(%rsp) # 8-byte Folded Spill adcq $0, 152(%rsp) # 8-byte Folded Spill adcq $0, 112(%rsp) # 8-byte Folded Spill adcq $0, %r12 movq %rax, %rbx movq %rbx, %rdx imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 536(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 536(%rsp), %rbx movq 96(%rsp), %rax # 8-byte Reload adcq 544(%rsp), %rax movq 136(%rsp), %rcx # 8-byte Reload adcq 552(%rsp), %rcx movq %rcx, 136(%rsp) # 8-byte Spill movq 168(%rsp), %rcx # 8-byte Reload adcq 560(%rsp), %rcx movq %rcx, 168(%rsp) # 8-byte Spill adcq 568(%rsp), %r15 movq %r15, 176(%rsp) # 8-byte Spill adcq 576(%rsp), %r14 movq %r14, 192(%rsp) # 8-byte Spill adcq 584(%rsp), %rbp movq %rbp, 160(%rsp) # 8-byte Spill movq 200(%rsp), %r13 # 8-byte Reload adcq 592(%rsp), %r13 movq 208(%rsp), %r15 # 8-byte Reload adcq 600(%rsp), %r15 movq 184(%rsp), %rbp # 8-byte Reload adcq 608(%rsp), %rbp movq 144(%rsp), %rbx # 8-byte Reload adcq $0, %rbx adcq $0, 104(%rsp) # 8-byte Folded Spill adcq $0, 152(%rsp) # 8-byte Folded Spill adcq $0, 112(%rsp) # 8-byte Folded Spill adcq $0, %r12 movq %rax, %rdx movq %rax, %r14 imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 456(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 456(%rsp), %r14 movq 136(%rsp), %rax # 8-byte Reload adcq 464(%rsp), %rax movq 168(%rsp), %rcx # 8-byte Reload adcq 472(%rsp), %rcx movq %rcx, 168(%rsp) # 8-byte Spill movq 176(%rsp), %rcx # 8-byte Reload adcq 480(%rsp), %rcx movq %rcx, 176(%rsp) # 8-byte Spill movq 192(%rsp), %rcx # 8-byte Reload adcq 488(%rsp), %rcx movq %rcx, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rcx # 8-byte Reload adcq 496(%rsp), %rcx movq %rcx, 160(%rsp) # 8-byte Spill adcq 504(%rsp), %r13 movq %r13, 200(%rsp) # 8-byte Spill adcq 512(%rsp), %r15 movq %r15, 208(%rsp) # 8-byte Spill adcq 520(%rsp), %rbp movq %rbp, 184(%rsp) # 8-byte Spill adcq 528(%rsp), %rbx movq %rbx, 144(%rsp) # 8-byte Spill movq 104(%rsp), %r14 # 8-byte Reload adcq $0, %r14 movq 152(%rsp), %r13 # 8-byte Reload adcq $0, %r13 movq 112(%rsp), %rbx # 8-byte Reload adcq $0, %rbx adcq $0, %r12 movq %rax, %rdx movq %rax, %r15 imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 376(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 376(%rsp), %r15 movq 168(%rsp), %rax # 8-byte Reload adcq 384(%rsp), %rax movq 176(%rsp), %rcx # 8-byte Reload adcq 392(%rsp), %rcx movq %rcx, 176(%rsp) # 8-byte Spill movq 192(%rsp), %rcx # 8-byte Reload adcq 400(%rsp), %rcx movq %rcx, 192(%rsp) # 8-byte Spill movq 160(%rsp), %rbp # 8-byte Reload adcq 408(%rsp), %rbp movq 200(%rsp), %rcx # 8-byte Reload adcq 416(%rsp), %rcx movq %rcx, 200(%rsp) # 8-byte Spill movq 208(%rsp), %rcx # 8-byte Reload adcq 424(%rsp), %rcx movq %rcx, 208(%rsp) # 8-byte Spill movq 184(%rsp), %rcx # 8-byte Reload adcq 432(%rsp), %rcx movq %rcx, 184(%rsp) # 8-byte Spill movq 144(%rsp), %r15 # 8-byte Reload adcq 440(%rsp), %r15 adcq 448(%rsp), %r14 movq %r14, 104(%rsp) # 8-byte Spill adcq $0, %r13 movq %r13, %r14 adcq $0, %rbx movq %rbx, 112(%rsp) # 8-byte Spill adcq $0, %r12 movq %rax, %rbx movq %rbx, %rdx imulq 120(%rsp), %rdx # 8-byte Folded Reload leaq 296(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 296(%rsp), %rbx movq 176(%rsp), %rax # 8-byte Reload adcq 304(%rsp), %rax movq 192(%rsp), %r13 # 8-byte Reload adcq 312(%rsp), %r13 adcq 320(%rsp), %rbp movq 200(%rsp), %rcx # 8-byte Reload adcq 328(%rsp), %rcx movq %rcx, 200(%rsp) # 8-byte Spill movq 208(%rsp), %rcx # 8-byte Reload adcq 336(%rsp), %rcx movq %rcx, 208(%rsp) # 8-byte Spill movq 184(%rsp), %rcx # 8-byte Reload adcq 344(%rsp), %rcx movq %rcx, 184(%rsp) # 8-byte Spill adcq 352(%rsp), %r15 movq %r15, 144(%rsp) # 8-byte Spill movq 104(%rsp), %r15 # 8-byte Reload adcq 360(%rsp), %r15 adcq 368(%rsp), %r14 movq %r14, 152(%rsp) # 8-byte Spill movq 112(%rsp), %r14 # 8-byte Reload adcq $0, %r14 adcq $0, %r12 movq 120(%rsp), %rdx # 8-byte Reload imulq %rax, %rdx movq %rax, %rbx leaq 216(%rsp), %rdi movq 128(%rsp), %rsi # 8-byte Reload callq .LmulPv576x64 addq 216(%rsp), %rbx movq %r13, %rsi adcq 224(%rsp), %rsi movq %rsi, 192(%rsp) # 8-byte Spill adcq 232(%rsp), %rbp movq %rbp, 160(%rsp) # 8-byte Spill movq 200(%rsp), %r9 # 8-byte Reload adcq 240(%rsp), %r9 movq %r9, 200(%rsp) # 8-byte Spill movq 208(%rsp), %r8 # 8-byte Reload adcq 248(%rsp), %r8 movq %r8, 208(%rsp) # 8-byte Spill movq 184(%rsp), %rbx # 8-byte Reload adcq 256(%rsp), %rbx movq 144(%rsp), %rax # 8-byte Reload adcq 264(%rsp), %rax movq %r15, %rcx adcq 272(%rsp), %rcx movq 152(%rsp), %rdx # 8-byte Reload adcq 280(%rsp), %rdx movq %rdx, 152(%rsp) # 8-byte Spill adcq 288(%rsp), %r14 movq %r14, %r11 adcq $0, %r12 subq 16(%rsp), %rsi # 8-byte Folded Reload movq %rbp, %rdi sbbq 8(%rsp), %rdi # 8-byte Folded Reload movq %r9, %rbp sbbq 24(%rsp), %rbp # 8-byte Folded Reload movq %r8, %r13 sbbq 32(%rsp), %r13 # 8-byte Folded Reload movq %rbx, %r15 sbbq 40(%rsp), %r15 # 8-byte Folded Reload movq %rax, %r14 sbbq 48(%rsp), %r14 # 8-byte Folded Reload movq %rcx, %r10 sbbq 56(%rsp), %r10 # 8-byte Folded Reload movq %rdx, %r8 sbbq 64(%rsp), %r8 # 8-byte Folded Reload movq %r11, %r9 sbbq 72(%rsp), %r9 # 8-byte Folded Reload sbbq $0, %r12 andl $1, %r12d cmovneq %r11, %r9 testb %r12b, %r12b cmovneq 192(%rsp), %rsi # 8-byte Folded Reload movq 80(%rsp), %rdx # 8-byte Reload movq %rsi, (%rdx) cmovneq 160(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 8(%rdx) cmovneq 200(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 16(%rdx) cmovneq 208(%rsp), %r13 # 8-byte Folded Reload movq %r13, 24(%rdx) cmovneq %rbx, %r15 movq %r15, 32(%rdx) cmovneq %rax, %r14 movq %r14, 40(%rdx) cmovneq %rcx, %r10 movq %r10, 48(%rdx) cmovneq 152(%rsp), %r8 # 8-byte Folded Reload movq %r8, 56(%rdx) movq %r9, 64(%rdx) addq $936, %rsp # imm = 0x3A8 popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end132: .size mcl_fp_montRed9L, .Lfunc_end132-mcl_fp_montRed9L .globl mcl_fp_addPre9L .align 16, 0x90 .type mcl_fp_addPre9L,@function mcl_fp_addPre9L: # @mcl_fp_addPre9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 64(%rdx), %r8 movq 64(%rsi), %r15 movq 56(%rsi), %r9 movq 48(%rsi), %r10 movq 40(%rsi), %r11 movq 24(%rsi), %r12 movq 32(%rsi), %r14 movq (%rdx), %rbx movq 8(%rdx), %rcx addq (%rsi), %rbx adcq 8(%rsi), %rcx movq 16(%rdx), %rax adcq 16(%rsi), %rax adcq 24(%rdx), %r12 movq 56(%rdx), %r13 movq 48(%rdx), %rsi movq 40(%rdx), %rbp movq 32(%rdx), %rdx movq %rbx, (%rdi) movq %rcx, 8(%rdi) movq %rax, 16(%rdi) movq %r12, 24(%rdi) adcq %r14, %rdx movq %rdx, 32(%rdi) adcq %r11, %rbp movq %rbp, 40(%rdi) adcq %r10, %rsi movq %rsi, 48(%rdi) adcq %r9, %r13 movq %r13, 56(%rdi) adcq %r8, %r15 movq %r15, 64(%rdi) sbbq %rax, %rax andl $1, %eax popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end133: .size mcl_fp_addPre9L, .Lfunc_end133-mcl_fp_addPre9L .globl mcl_fp_subPre9L .align 16, 0x90 .type mcl_fp_subPre9L,@function mcl_fp_subPre9L: # @mcl_fp_subPre9L # BB#0: movq 32(%rdx), %r8 movq (%rsi), %rcx xorl %eax, %eax subq (%rdx), %rcx movq %rcx, (%rdi) movq 8(%rsi), %rcx sbbq 8(%rdx), %rcx movq %rcx, 8(%rdi) movq 16(%rsi), %rcx sbbq 16(%rdx), %rcx movq %rcx, 16(%rdi) movq 24(%rsi), %rcx sbbq 24(%rdx), %rcx movq %rcx, 24(%rdi) movq 32(%rsi), %rcx sbbq %r8, %rcx movq 40(%rdx), %r8 movq %rcx, 32(%rdi) movq 40(%rsi), %rcx sbbq %r8, %rcx movq 48(%rdx), %r8 movq %rcx, 40(%rdi) movq 48(%rsi), %rcx sbbq %r8, %rcx movq 56(%rdx), %r8 movq %rcx, 48(%rdi) movq 56(%rsi), %rcx sbbq %r8, %rcx movq %rcx, 56(%rdi) movq 64(%rdx), %rcx movq 64(%rsi), %rdx sbbq %rcx, %rdx movq %rdx, 64(%rdi) sbbq $0, %rax andl $1, %eax retq .Lfunc_end134: .size mcl_fp_subPre9L, .Lfunc_end134-mcl_fp_subPre9L .globl mcl_fp_shr1_9L .align 16, 0x90 .type mcl_fp_shr1_9L,@function mcl_fp_shr1_9L: # @mcl_fp_shr1_9L # BB#0: pushq %rbx movq 64(%rsi), %r8 movq 56(%rsi), %r9 movq 48(%rsi), %r10 movq 40(%rsi), %r11 movq 32(%rsi), %rcx movq 24(%rsi), %rdx movq 16(%rsi), %rax movq (%rsi), %rbx movq 8(%rsi), %rsi shrdq $1, %rsi, %rbx movq %rbx, (%rdi) shrdq $1, %rax, %rsi movq %rsi, 8(%rdi) shrdq $1, %rdx, %rax movq %rax, 16(%rdi) shrdq $1, %rcx, %rdx movq %rdx, 24(%rdi) shrdq $1, %r11, %rcx movq %rcx, 32(%rdi) shrdq $1, %r10, %r11 movq %r11, 40(%rdi) shrdq $1, %r9, %r10 movq %r10, 48(%rdi) shrdq $1, %r8, %r9 movq %r9, 56(%rdi) shrq %r8 movq %r8, 64(%rdi) popq %rbx retq .Lfunc_end135: .size mcl_fp_shr1_9L, .Lfunc_end135-mcl_fp_shr1_9L .globl mcl_fp_add9L .align 16, 0x90 .type mcl_fp_add9L,@function mcl_fp_add9L: # @mcl_fp_add9L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 64(%rdx), %r12 movq 64(%rsi), %r8 movq 56(%rsi), %r13 movq 48(%rsi), %r9 movq 40(%rsi), %r10 movq 24(%rsi), %r14 movq 32(%rsi), %r11 movq (%rdx), %rbx movq 8(%rdx), %r15 addq (%rsi), %rbx adcq 8(%rsi), %r15 movq 16(%rdx), %rax adcq 16(%rsi), %rax adcq 24(%rdx), %r14 adcq 32(%rdx), %r11 adcq 40(%rdx), %r10 movq 56(%rdx), %rsi adcq 48(%rdx), %r9 movq %rbx, (%rdi) movq %r15, 8(%rdi) movq %rax, 16(%rdi) movq %r14, 24(%rdi) movq %r11, 32(%rdi) movq %r10, 40(%rdi) movq %r9, 48(%rdi) adcq %r13, %rsi movq %rsi, 56(%rdi) adcq %r12, %r8 movq %r8, 64(%rdi) sbbq %rdx, %rdx andl $1, %edx subq (%rcx), %rbx sbbq 8(%rcx), %r15 sbbq 16(%rcx), %rax sbbq 24(%rcx), %r14 sbbq 32(%rcx), %r11 sbbq 40(%rcx), %r10 sbbq 48(%rcx), %r9 sbbq 56(%rcx), %rsi sbbq 64(%rcx), %r8 sbbq $0, %rdx testb $1, %dl jne .LBB136_2 # BB#1: # %nocarry movq %rbx, (%rdi) movq %r15, 8(%rdi) movq %rax, 16(%rdi) movq %r14, 24(%rdi) movq %r11, 32(%rdi) movq %r10, 40(%rdi) movq %r9, 48(%rdi) movq %rsi, 56(%rdi) movq %r8, 64(%rdi) .LBB136_2: # %carry popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end136: .size mcl_fp_add9L, .Lfunc_end136-mcl_fp_add9L .globl mcl_fp_addNF9L .align 16, 0x90 .type mcl_fp_addNF9L,@function mcl_fp_addNF9L: # @mcl_fp_addNF9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rdi, %r8 movq 64(%rdx), %r10 movq 56(%rdx), %r11 movq 48(%rdx), %r9 movq 40(%rdx), %rax movq 32(%rdx), %rdi movq 24(%rdx), %rbp movq 16(%rdx), %r15 movq (%rdx), %rbx movq 8(%rdx), %r13 addq (%rsi), %rbx adcq 8(%rsi), %r13 adcq 16(%rsi), %r15 adcq 24(%rsi), %rbp movq %rbp, -40(%rsp) # 8-byte Spill adcq 32(%rsi), %rdi movq %rdi, -16(%rsp) # 8-byte Spill adcq 40(%rsi), %rax movq %rax, -8(%rsp) # 8-byte Spill adcq 48(%rsi), %r9 movq %r9, -32(%rsp) # 8-byte Spill movq %r9, %rdi adcq 56(%rsi), %r11 movq %r11, -24(%rsp) # 8-byte Spill movq %r11, %rax adcq 64(%rsi), %r10 movq %r10, %r9 movq %rbx, %rsi subq (%rcx), %rsi movq %r13, %rdx sbbq 8(%rcx), %rdx movq %r15, %r12 sbbq 16(%rcx), %r12 sbbq 24(%rcx), %rbp movq -16(%rsp), %r14 # 8-byte Reload sbbq 32(%rcx), %r14 movq -8(%rsp), %r11 # 8-byte Reload sbbq 40(%rcx), %r11 movq %rdi, %r10 sbbq 48(%rcx), %r10 movq %rax, %rdi sbbq 56(%rcx), %rdi movq %r9, %rax sbbq 64(%rcx), %rax movq %rax, %rcx sarq $63, %rcx cmovsq %rbx, %rsi movq %rsi, (%r8) cmovsq %r13, %rdx movq %rdx, 8(%r8) cmovsq %r15, %r12 movq %r12, 16(%r8) cmovsq -40(%rsp), %rbp # 8-byte Folded Reload movq %rbp, 24(%r8) cmovsq -16(%rsp), %r14 # 8-byte Folded Reload movq %r14, 32(%r8) cmovsq -8(%rsp), %r11 # 8-byte Folded Reload movq %r11, 40(%r8) cmovsq -32(%rsp), %r10 # 8-byte Folded Reload movq %r10, 48(%r8) cmovsq -24(%rsp), %rdi # 8-byte Folded Reload movq %rdi, 56(%r8) cmovsq %r9, %rax movq %rax, 64(%r8) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end137: .size mcl_fp_addNF9L, .Lfunc_end137-mcl_fp_addNF9L .globl mcl_fp_sub9L .align 16, 0x90 .type mcl_fp_sub9L,@function mcl_fp_sub9L: # @mcl_fp_sub9L # BB#0: pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq 64(%rdx), %r13 movq (%rsi), %rax movq 8(%rsi), %r9 xorl %ebx, %ebx subq (%rdx), %rax sbbq 8(%rdx), %r9 movq 16(%rsi), %r10 sbbq 16(%rdx), %r10 movq 24(%rsi), %r11 sbbq 24(%rdx), %r11 movq 32(%rsi), %r12 sbbq 32(%rdx), %r12 movq 40(%rsi), %r14 sbbq 40(%rdx), %r14 movq 48(%rsi), %r15 sbbq 48(%rdx), %r15 movq 64(%rsi), %r8 movq 56(%rsi), %rsi sbbq 56(%rdx), %rsi movq %rax, (%rdi) movq %r9, 8(%rdi) movq %r10, 16(%rdi) movq %r11, 24(%rdi) movq %r12, 32(%rdi) movq %r14, 40(%rdi) movq %r15, 48(%rdi) movq %rsi, 56(%rdi) sbbq %r13, %r8 movq %r8, 64(%rdi) sbbq $0, %rbx testb $1, %bl je .LBB138_2 # BB#1: # %carry addq (%rcx), %rax movq %rax, (%rdi) movq 8(%rcx), %rax adcq %r9, %rax movq %rax, 8(%rdi) movq 16(%rcx), %rax adcq %r10, %rax movq %rax, 16(%rdi) movq 24(%rcx), %rax adcq %r11, %rax movq %rax, 24(%rdi) movq 32(%rcx), %rax adcq %r12, %rax movq %rax, 32(%rdi) movq 40(%rcx), %rax adcq %r14, %rax movq %rax, 40(%rdi) movq 48(%rcx), %rax adcq %r15, %rax movq %rax, 48(%rdi) movq 56(%rcx), %rax adcq %rsi, %rax movq %rax, 56(%rdi) movq 64(%rcx), %rax adcq %r8, %rax movq %rax, 64(%rdi) .LBB138_2: # %nocarry popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 retq .Lfunc_end138: .size mcl_fp_sub9L, .Lfunc_end138-mcl_fp_sub9L .globl mcl_fp_subNF9L .align 16, 0x90 .type mcl_fp_subNF9L,@function mcl_fp_subNF9L: # @mcl_fp_subNF9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r8 movq %rdi, %r11 movq 64(%rsi), %r14 movq 56(%rsi), %rax movq 48(%rsi), %rcx movq 40(%rsi), %rdi movq 32(%rsi), %rbp movq 24(%rsi), %rbx movq 16(%rsi), %r15 movq (%rsi), %r12 movq 8(%rsi), %r13 subq (%rdx), %r12 sbbq 8(%rdx), %r13 sbbq 16(%rdx), %r15 sbbq 24(%rdx), %rbx movq %rbx, -40(%rsp) # 8-byte Spill sbbq 32(%rdx), %rbp movq %rbp, -32(%rsp) # 8-byte Spill sbbq 40(%rdx), %rdi movq %rdi, -24(%rsp) # 8-byte Spill sbbq 48(%rdx), %rcx movq %rcx, -16(%rsp) # 8-byte Spill sbbq 56(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill sbbq 64(%rdx), %r14 movq %r14, %rax sarq $63, %rax movq %rax, %rcx shldq $1, %r14, %rcx movq 24(%r8), %rbp andq %rcx, %rbp movq 8(%r8), %rdi andq %rcx, %rdi andq (%r8), %rcx movq 64(%r8), %rbx andq %rax, %rbx movq 56(%r8), %r10 andq %rax, %r10 rolq %rax movq 48(%r8), %r9 andq %rax, %r9 movq 40(%r8), %rsi andq %rax, %rsi movq 32(%r8), %rdx andq %rax, %rdx andq 16(%r8), %rax addq %r12, %rcx adcq %r13, %rdi movq %rcx, (%r11) adcq %r15, %rax movq %rdi, 8(%r11) adcq -40(%rsp), %rbp # 8-byte Folded Reload movq %rax, 16(%r11) movq %rbp, 24(%r11) adcq -32(%rsp), %rdx # 8-byte Folded Reload movq %rdx, 32(%r11) adcq -24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 40(%r11) adcq -16(%rsp), %r9 # 8-byte Folded Reload movq %r9, 48(%r11) adcq -8(%rsp), %r10 # 8-byte Folded Reload movq %r10, 56(%r11) adcq %r14, %rbx movq %rbx, 64(%r11) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end139: .size mcl_fp_subNF9L, .Lfunc_end139-mcl_fp_subNF9L .globl mcl_fpDbl_add9L .align 16, 0x90 .type mcl_fpDbl_add9L,@function mcl_fpDbl_add9L: # @mcl_fpDbl_add9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r15 movq 136(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 128(%rdx), %rax movq %rax, -40(%rsp) # 8-byte Spill movq 120(%rdx), %r10 movq 112(%rdx), %r11 movq 24(%rsi), %rcx movq 32(%rsi), %r14 movq 16(%rdx), %rbp movq (%rdx), %rax movq 8(%rdx), %rbx addq (%rsi), %rax adcq 8(%rsi), %rbx adcq 16(%rsi), %rbp adcq 24(%rdx), %rcx adcq 32(%rdx), %r14 movq 104(%rdx), %r9 movq 96(%rdx), %r13 movq %rax, (%rdi) movq 88(%rdx), %r8 movq %rbx, 8(%rdi) movq 80(%rdx), %r12 movq %rbp, 16(%rdi) movq 40(%rdx), %rax movq %rcx, 24(%rdi) movq 40(%rsi), %rbp adcq %rax, %rbp movq 48(%rdx), %rcx movq %r14, 32(%rdi) movq 48(%rsi), %rax adcq %rcx, %rax movq 56(%rdx), %r14 movq %rbp, 40(%rdi) movq 56(%rsi), %rbp adcq %r14, %rbp movq 72(%rdx), %rcx movq 64(%rdx), %rdx movq %rax, 48(%rdi) movq 64(%rsi), %rax adcq %rdx, %rax movq 136(%rsi), %rbx movq %rbp, 56(%rdi) movq 72(%rsi), %rbp adcq %rcx, %rbp movq 128(%rsi), %rcx movq %rax, 64(%rdi) movq 80(%rsi), %rdx adcq %r12, %rdx movq 88(%rsi), %r12 adcq %r8, %r12 movq 96(%rsi), %r14 adcq %r13, %r14 movq %r14, -48(%rsp) # 8-byte Spill movq 104(%rsi), %rax adcq %r9, %rax movq %rax, -32(%rsp) # 8-byte Spill movq 120(%rsi), %rax movq 112(%rsi), %rsi adcq %r11, %rsi movq %rsi, -24(%rsp) # 8-byte Spill adcq %r10, %rax movq %rax, -16(%rsp) # 8-byte Spill adcq -40(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -40(%rsp) # 8-byte Spill adcq -8(%rsp), %rbx # 8-byte Folded Reload movq %rbx, -8(%rsp) # 8-byte Spill sbbq %r9, %r9 andl $1, %r9d movq %rbp, %r10 subq (%r15), %r10 movq %rdx, %r11 sbbq 8(%r15), %r11 movq %r12, %rbx sbbq 16(%r15), %rbx sbbq 24(%r15), %r14 movq -32(%rsp), %r13 # 8-byte Reload sbbq 32(%r15), %r13 movq -24(%rsp), %rsi # 8-byte Reload sbbq 40(%r15), %rsi movq -16(%rsp), %rax # 8-byte Reload sbbq 48(%r15), %rax sbbq 56(%r15), %rcx movq -8(%rsp), %r8 # 8-byte Reload sbbq 64(%r15), %r8 sbbq $0, %r9 andl $1, %r9d cmovneq %rbp, %r10 movq %r10, 72(%rdi) testb %r9b, %r9b cmovneq %rdx, %r11 movq %r11, 80(%rdi) cmovneq %r12, %rbx movq %rbx, 88(%rdi) cmovneq -48(%rsp), %r14 # 8-byte Folded Reload movq %r14, 96(%rdi) cmovneq -32(%rsp), %r13 # 8-byte Folded Reload movq %r13, 104(%rdi) cmovneq -24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 112(%rdi) cmovneq -16(%rsp), %rax # 8-byte Folded Reload movq %rax, 120(%rdi) cmovneq -40(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 128(%rdi) cmovneq -8(%rsp), %r8 # 8-byte Folded Reload movq %r8, 136(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end140: .size mcl_fpDbl_add9L, .Lfunc_end140-mcl_fpDbl_add9L .globl mcl_fpDbl_sub9L .align 16, 0x90 .type mcl_fpDbl_sub9L,@function mcl_fpDbl_sub9L: # @mcl_fpDbl_sub9L # BB#0: pushq %rbp pushq %r15 pushq %r14 pushq %r13 pushq %r12 pushq %rbx movq %rcx, %r14 movq 136(%rdx), %rax movq %rax, -8(%rsp) # 8-byte Spill movq 128(%rdx), %rax movq %rax, -16(%rsp) # 8-byte Spill movq 120(%rdx), %rax movq %rax, -24(%rsp) # 8-byte Spill movq 16(%rsi), %r11 movq (%rsi), %r12 movq 8(%rsi), %r13 xorl %r9d, %r9d subq (%rdx), %r12 sbbq 8(%rdx), %r13 sbbq 16(%rdx), %r11 movq 24(%rsi), %rbx sbbq 24(%rdx), %rbx movq 32(%rsi), %rbp sbbq 32(%rdx), %rbp movq 112(%rdx), %r10 movq 104(%rdx), %rcx movq %r12, (%rdi) movq 96(%rdx), %rax movq %r13, 8(%rdi) movq 88(%rdx), %r13 movq %r11, 16(%rdi) movq 40(%rdx), %r11 movq %rbx, 24(%rdi) movq 40(%rsi), %rbx sbbq %r11, %rbx movq 48(%rdx), %r11 movq %rbp, 32(%rdi) movq 48(%rsi), %rbp sbbq %r11, %rbp movq 56(%rdx), %r11 movq %rbx, 40(%rdi) movq 56(%rsi), %rbx sbbq %r11, %rbx movq 64(%rdx), %r11 movq %rbp, 48(%rdi) movq 64(%rsi), %rbp sbbq %r11, %rbp movq 80(%rdx), %r8 movq 72(%rdx), %r11 movq %rbx, 56(%rdi) movq 72(%rsi), %r15 sbbq %r11, %r15 movq 136(%rsi), %rdx movq %rbp, 64(%rdi) movq 80(%rsi), %rbp sbbq %r8, %rbp movq 88(%rsi), %r12 sbbq %r13, %r12 movq 96(%rsi), %r13 sbbq %rax, %r13 movq 104(%rsi), %rax sbbq %rcx, %rax movq %rax, -40(%rsp) # 8-byte Spill movq 112(%rsi), %rax sbbq %r10, %rax movq %rax, -32(%rsp) # 8-byte Spill movq 128(%rsi), %rax movq 120(%rsi), %rcx sbbq -24(%rsp), %rcx # 8-byte Folded Reload movq %rcx, -24(%rsp) # 8-byte Spill sbbq -16(%rsp), %rax # 8-byte Folded Reload movq %rax, -16(%rsp) # 8-byte Spill sbbq -8(%rsp), %rdx # 8-byte Folded Reload movq %rdx, -8(%rsp) # 8-byte Spill movl $0, %r8d sbbq $0, %r8 andl $1, %r8d movq (%r14), %r10 cmoveq %r9, %r10 testb %r8b, %r8b movq 16(%r14), %r8 cmoveq %r9, %r8 movq 8(%r14), %rdx cmoveq %r9, %rdx movq 64(%r14), %rbx cmoveq %r9, %rbx movq 56(%r14), %r11 cmoveq %r9, %r11 movq 48(%r14), %rsi cmoveq %r9, %rsi movq 40(%r14), %rcx cmoveq %r9, %rcx movq 32(%r14), %rax cmoveq %r9, %rax cmovneq 24(%r14), %r9 addq %r15, %r10 adcq %rbp, %rdx movq %r10, 72(%rdi) adcq %r12, %r8 movq %rdx, 80(%rdi) adcq %r13, %r9 movq %r8, 88(%rdi) movq %r9, 96(%rdi) adcq -40(%rsp), %rax # 8-byte Folded Reload movq %rax, 104(%rdi) adcq -32(%rsp), %rcx # 8-byte Folded Reload movq %rcx, 112(%rdi) adcq -24(%rsp), %rsi # 8-byte Folded Reload movq %rsi, 120(%rdi) adcq -16(%rsp), %r11 # 8-byte Folded Reload movq %r11, 128(%rdi) adcq -8(%rsp), %rbx # 8-byte Folded Reload movq %rbx, 136(%rdi) popq %rbx popq %r12 popq %r13 popq %r14 popq %r15 popq %rbp retq .Lfunc_end141: .size mcl_fpDbl_sub9L, .Lfunc_end141-mcl_fpDbl_sub9L .section ".note.GNU-stack","",@progbits