/*---------------------------------------------------------------------------+ | mul_Xsig.S | | | | Multiply a 12 byte fixed point number by another fixed point number. | | | | Copyright (C) 1992,1994 | | W. Metzenthen, 22 Parker St, Ormond, Vic 3163, | | Australia. E-mail billm@vaxc.cc.monash.edu.au | | | | Call from C as: | | void mul32_Xsig(Xsig *x, unsigned b) | | | | void mul64_Xsig(Xsig *x, unsigned long long *b) | | | | void mul_Xsig_Xsig(Xsig *x, unsigned *b) | | | | The result is neither rounded nor normalized, and the ls bit or so may | | be wrong. | | | +---------------------------------------------------------------------------*/ .file "mul_Xsig.S" #include "fpu_asm.h" .text ENTRY(mul32_Xsig) pushl %ebp movl %esp,%ebp subl $16,%esp pushl %esi movl PARAM1,%esi movl PARAM2,%ecx xor %eax,%eax movl %eax,-4(%ebp) movl %eax,-8(%ebp) movl (%esi),%eax /* lsl of Xsig */ mull %ecx /* msl of b */ movl %edx,-12(%ebp) movl 4(%esi),%eax /* midl of Xsig */ mull %ecx /* msl of b */ addl %eax,-12(%ebp) adcl %edx,-8(%ebp) adcl $0,-4(%ebp) movl 8(%esi),%eax /* msl of Xsig */ mull %ecx /* msl of b */ addl %eax,-8(%ebp) adcl %edx,-4(%ebp) movl -12(%ebp),%eax movl %eax,(%esi) movl -8(%ebp),%eax movl %eax,4(%esi) movl -4(%ebp),%eax movl %eax,8(%esi) popl %esi leave ret ENTRY(mul64_Xsig) pushl %ebp movl %esp,%ebp subl $16,%esp pushl %esi movl PARAM1,%esi movl PARAM2,%ecx xor %eax,%eax movl %eax,-4(%ebp) movl %eax,-8(%ebp) movl (%esi),%eax /* lsl of Xsig */ mull 4(%ecx) /* msl of b */ movl %edx,-12(%ebp) movl 4(%esi),%eax /* midl of Xsig */ mull (%ecx) /* lsl of b */ addl %edx,-12(%ebp) adcl $0,-8(%ebp) adcl $0,-4(%ebp) movl 4(%esi),%eax /* midl of Xsig */ mull 4(%ecx) /* msl of b */ addl %eax,-12(%ebp) adcl %edx,-8(%ebp) adcl $0,-4(%ebp) movl 8(%esi),%eax /* msl of Xsig */ mull (%ecx) /* lsl of b */ addl %eax,-12(%ebp) adcl %edx,-8(%ebp) adcl $0,-4(%ebp) movl 8(%esi),%eax /* msl of Xsig */ mull 4(%ecx) /* msl of b */ addl %eax,-8(%ebp) adcl %edx,-4(%ebp) movl -12(%ebp),%eax movl %eax,(%esi) movl -8(%ebp),%eax movl %eax,4(%esi) movl -4(%ebp),%eax movl %eax,8(%esi) popl %esi leave ret ENTRY(mul_Xsig_Xsig) pushl %ebp movl %esp,%ebp subl $16,%esp pushl %esi movl PARAM1,%esi movl PARAM2,%ecx xor %eax,%eax movl %eax,-4(%ebp) movl %eax,-8(%ebp) movl (%esi),%eax /* lsl of Xsig */ mull 8(%ecx) /* msl of b */ movl %edx,-12(%ebp) movl 4(%esi),%eax /* midl of Xsig */ mull 4(%ecx) /* midl of b */ addl %edx,-12(%ebp) adcl $0,-8(%ebp) adcl $0,-4(%ebp) movl 8(%esi),%eax /* msl of Xsig */ mull (%ecx) /* lsl of b */ addl %edx,-12(%ebp) adcl $0,-8(%ebp) adcl $0,-4(%ebp) movl 4(%esi),%eax /* midl of Xsig */ mull 8(%ecx) /* msl of b */ addl %eax,-12(%ebp) adcl %edx,-8(%ebp) adcl $0,-4(%ebp) movl 8(%esi),%eax /* msl of Xsig */ mull 4(%ecx) /* midl of b */ addl %eax,-12(%ebp) adcl %edx,-8(%ebp) adcl $0,-4(%ebp) movl 8(%esi),%eax /* msl of Xsig */ mull 8(%ecx) /* msl of b */ addl %eax,-8(%ebp) adcl %edx,-4(%ebp) movl -12(%ebp),%edx movl %edx,(%esi) movl -8(%ebp),%edx movl %edx,4(%esi) movl -4(%ebp),%edx movl %edx,8(%esi) popl %esi leave ret