dnl AMD64 mpn_addlshC_n, mpn_sublshC_n -- rp[] = up[] +- (vp[] << C), where dnl C is 1, 2, 3. Optimized for Pentium 4. dnl Contributed to the GNU project by Torbjorn Granlund. dnl Copyright 2008, 2010, 2011, 2012 Free Software Foundation, Inc. dnl This file is part of the GNU MP Library. dnl The GNU MP Library is free software; you can redistribute it and/or modify dnl it under the terms of the GNU Lesser General Public License as published dnl by the Free Software Foundation; either version 3 of the License, or (at dnl your option) any later version. dnl The GNU MP Library is distributed in the hope that it will be useful, but dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public dnl License for more details. dnl You should have received a copy of the GNU Lesser General Public License dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. C cycles/limb C AMD K8,K9 3.8 C AMD K10 3.8 C Intel P4 5.8 C Intel core2 4.75 C Intel corei 4.75 C Intel atom ? C VIA nano 4.75 C INPUT PARAMETERS define(`rp',`%rdi') define(`up',`%rsi') define(`vp',`%rdx') define(`n', `%rcx') define(M, eval(m4_lshift(1,LSH))) ABI_SUPPORT(DOS64) ABI_SUPPORT(STD64) ASM_START() TEXT ALIGN(16) PROLOGUE(func) FUNC_ENTRY(4) push %rbx push %r12 push %rbp mov (vp), %r9 shl $LSH, %r9 mov 4(vp), R32(%rbp) xor R32(%rbx), R32(%rbx) mov R32(n), R32(%rax) and $3, R32(%rax) jne L(n00) C n = 0, 4, 8, ... mov (up), %r8 mov 8(up), %r10 shr $RSH, R32(%rbp) ADDSUB %r9, %r8 mov 8(vp), %r9 lea (%rbp,%r9,M), %r9 setc R8(%rax) mov 12(vp), R32(%rbp) lea -16(rp), rp jmp L(L00) L(n00): cmp $2, R32(%rax) jnc L(n01) C n = 1, 5, 9, ... mov (up), %r11 lea -8(rp), rp shr $RSH, R32(%rbp) ADDSUB %r9, %r11 setc R8(%rbx) dec n jz L(1) C jump for n = 1 mov 8(up), %r8 mov 8(vp), %r9 lea (%rbp,%r9,M), %r9 mov 12(vp), R32(%rbp) lea 8(up), up lea 8(vp), vp jmp L(L01) L(n01): jne L(n10) C n = 2, 6, 10, ... mov (up), %r12 mov 8(up), %r11 shr $RSH, R32(%rbp) ADDSUB %r9, %r12 mov 8(vp), %r9 lea (%rbp,%r9,M), %r9 setc R8(%rax) mov 12(vp), R32(%rbp) lea 16(up), up lea 16(vp), vp jmp L(L10) L(n10): mov (up), %r10 mov 8(up), %r12 shr $RSH, R32(%rbp) ADDSUB %r9, %r10 mov 8(vp), %r9 lea (%rbp,%r9,M), %r9 setc R8(%rbx) mov 12(vp), R32(%rbp) lea -24(rp), rp lea -8(up), up lea -8(vp), vp jmp L(L11) L(c0): mov $1, R8(%rbx) jmp L(rc0) L(c1): mov $1, R8(%rax) jmp L(rc1) L(c2): mov $1, R8(%rbx) jmp L(rc2) ALIGN(16) L(top): mov (up), %r8 C not on critical path shr $RSH, R32(%rbp) ADDSUB %r9, %r11 C not on critical path mov (vp), %r9 lea (%rbp,%r9,M), %r9 setc R8(%rbx) C save carry out mov 4(vp), R32(%rbp) mov %r12, (rp) ADDSUB %rax, %r11 C apply previous carry out jc L(c0) C jump if ripple L(rc0): L(L01): mov 8(up), %r10 shr $RSH, R32(%rbp) ADDSUB %r9, %r8 mov 8(vp), %r9 lea (%rbp,%r9,M), %r9 setc R8(%rax) mov 12(vp), R32(%rbp) mov %r11, 8(rp) ADDSUB %rbx, %r8 jc L(c1) L(rc1): L(L00): mov 16(up), %r12 shr $RSH, R32(%rbp) ADDSUB %r9, %r10 mov 16(vp), %r9 lea (%rbp,%r9,M), %r9 setc R8(%rbx) mov 20(vp), R32(%rbp) mov %r8, 16(rp) ADDSUB %rax, %r10 jc L(c2) L(rc2): L(L11): mov 24(up), %r11 shr $RSH, R32(%rbp) ADDSUB %r9, %r12 mov 24(vp), %r9 lea (%rbp,%r9,M), %r9 lea 32(up), up lea 32(vp), vp setc R8(%rax) mov -4(vp), R32(%rbp) mov %r10, 24(rp) ADDSUB %rbx, %r12 jc L(c3) L(rc3): lea 32(rp), rp L(L10): sub $4, n ja L(top) L(end): shr $RSH, R32(%rbp) ADDSUB %r9, %r11 setc R8(%rbx) mov %r12, (rp) ADDSUB %rax, %r11 jnc L(1) mov $1, R8(%rbx) L(1): mov %r11, 8(rp) lea (%rbx,%rbp), R32(%rax) pop %rbp pop %r12 pop %rbx emms FUNC_EXIT() ret L(c3): mov $1, R8(%rax) jmp L(rc3) EPILOGUE() ASM_END()