dnl AMD K7 mpn_copyi -- copy limb vector, incrementing. dnl Copyright 1999, 2000, 2002, 2003 Free Software Foundation, Inc. dnl This file is part of the GNU MP Library. dnl dnl The GNU MP Library is free software; you can redistribute it and/or modify dnl it under the terms of either: dnl dnl * the GNU Lesser General Public License as published by the Free dnl Software Foundation; either version 3 of the License, or (at your dnl option) any later version. dnl dnl or dnl dnl * the GNU General Public License as published by the Free Software dnl Foundation; either version 2 of the License, or (at your option) any dnl later version. dnl dnl or both in parallel, as here. dnl dnl The GNU MP Library is distributed in the hope that it will be useful, but dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License dnl for more details. dnl dnl You should have received copies of the GNU General Public License and the dnl GNU Lesser General Public License along with the GNU MP Library. If not, dnl see https://www.gnu.org/licenses/. include(`../config.m4') C alignment dst/src, A=0mod8 N=4mod8 C A/A A/N N/A N/N C K7 0.75 1.0 1.0 0.75 C void mpn_copyi (mp_ptr dst, mp_srcptr src, mp_size_t size); C C Copy src,size to dst,size. C C This code at 0.75 or 1.0 c/l is always faster than a plain rep movsl at C 1.33 c/l. C C The K7 can do a 64-bit load and 64-bit store in one cycle (optimization C guile 22007 appendix B), so 0.5 c/l should be possible, however nothing C under 0.7 c/l is known. Apparently only two 32-bit stores can be done in C one cycle, so perhaps some scheduling is needed to ensure it's a C load+store in each cycle, not store+store. C C If both source and destination are unaligned then one limb is processed at C the start to make them aligned and so get 0.75 c/l, whereas if they'd been C used unaligned it would be 1.5 c/l. defframe(PARAM_SIZE,12) defframe(PARAM_SRC, 8) defframe(PARAM_DST, 4) dnl parameter space reused define(SAVE_EBX,`PARAM_SIZE') dnl minimum 5 since the unrolled code can't handle less than 5 deflit(UNROLL_THRESHOLD, 5) TEXT ALIGN(32) PROLOGUE(mpn_copyi) deflit(`FRAME',0) movl PARAM_SIZE, %ecx movl %ebx, SAVE_EBX movl PARAM_SRC, %eax movl PARAM_DST, %edx cmpl $UNROLL_THRESHOLD, %ecx jae L(unroll) orl %ecx, %ecx jz L(simple_done) L(simple): C eax src, incrementing C ebx scratch C ecx counter C edx dst, incrementing C C this loop is 2 cycles/limb movl (%eax), %ebx movl %ebx, (%edx) decl %ecx leal 4(%eax), %eax leal 4(%edx), %edx jnz L(simple) L(simple_done): movl SAVE_EBX, %ebx ret L(unroll): movl %eax, %ebx leal -12(%eax,%ecx,4), %eax C src end - 12 subl $3, %ecx C size-3 andl %edx, %ebx leal (%edx,%ecx,4), %edx C dst end - 12 negl %ecx testl $4, %ebx C testl to pad code closer to 16 bytes for L(top) jz L(aligned) C both src and dst unaligned, process one limb to align them movl (%eax,%ecx,4), %ebx movl %ebx, (%edx,%ecx,4) incl %ecx L(aligned): ALIGN(16) L(top): C eax src end - 12 C ebx C ecx counter, negative, limbs C edx dst end - 12 movq (%eax,%ecx,4), %mm0 movq 8(%eax,%ecx,4), %mm1 addl $4, %ecx movq %mm0, -16(%edx,%ecx,4) movq %mm1, -16+8(%edx,%ecx,4) ja L(top) C jump no carry and not zero C now %ecx is 0 to 3 representing respectively 3 to 0 limbs remaining testb $2, %cl jnz L(finish_not_two) movq (%eax,%ecx,4), %mm0 movq %mm0, (%edx,%ecx,4) L(finish_not_two): testb $1, %cl jnz L(done) movl 8(%eax), %ebx movl %ebx, 8(%edx) L(done): movl SAVE_EBX, %ebx emms ret EPILOGUE()