
This is achieved by filtering perlasm output through arm-xlate.pl. But note that it's done only if "flavour" argument is not 'void'. As 'void' is default value for other ARM targets, permasm output is not actually filtered on previously validated platforms. Reviewed-by: Dr. Stephen Henson <steve@openssl.org>
298 lines
7.1 KiB
Prolog
298 lines
7.1 KiB
Prolog
#!/usr/bin/env perl
|
||
#
|
||
# ====================================================================
|
||
# Written by Andy Polyakov <appro@openssl.org> for the OpenSSL
|
||
# project. The module is, however, dual licensed under OpenSSL and
|
||
# CRYPTOGAMS licenses depending on where you obtain it. For further
|
||
# details see http://www.openssl.org/~appro/cryptogams/.
|
||
# ====================================================================
|
||
#
|
||
# May 2011
|
||
#
|
||
# The module implements bn_GF2m_mul_2x2 polynomial multiplication
|
||
# used in bn_gf2m.c. It's kind of low-hanging mechanical port from
|
||
# C for the time being... Except that it has two code paths: pure
|
||
# integer code suitable for any ARMv4 and later CPU and NEON code
|
||
# suitable for ARMv7. Pure integer 1x1 multiplication subroutine runs
|
||
# in ~45 cycles on dual-issue core such as Cortex A8, which is ~50%
|
||
# faster than compiler-generated code. For ECDH and ECDSA verify (but
|
||
# not for ECDSA sign) it means 25%-45% improvement depending on key
|
||
# length, more for longer keys. Even though NEON 1x1 multiplication
|
||
# runs in even less cycles, ~30, improvement is measurable only on
|
||
# longer keys. One has to optimize code elsewhere to get NEON glow...
|
||
|
||
$flavour = shift;
|
||
if ($flavour=~/^\w[\w\-]*\.\w+$/) { $output=$flavour; undef $flavour; }
|
||
else { while (($output=shift) && ($output!~/^\w[\w\-]*\.\w+$/)) {} }
|
||
|
||
if ($flavour && $flavour ne "void") {
|
||
$0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
|
||
( $xlate="${dir}arm-xlate.pl" and -f $xlate ) or
|
||
( $xlate="${dir}../../perlasm/arm-xlate.pl" and -f $xlate) or
|
||
die "can't locate arm-xlate.pl";
|
||
|
||
open STDOUT,"| \"$^X\" $xlate $flavour $output";
|
||
} else {
|
||
open STDOUT,">$output";
|
||
}
|
||
|
||
sub Dlo() { shift=~m|q([1]?[0-9])|?"d".($1*2):""; }
|
||
sub Dhi() { shift=~m|q([1]?[0-9])|?"d".($1*2+1):""; }
|
||
sub Q() { shift=~m|d([1-3]?[02468])|?"q".($1/2):""; }
|
||
|
||
$code=<<___;
|
||
#include "arm_arch.h"
|
||
|
||
.text
|
||
.code 32
|
||
|
||
#if __ARM_ARCH__>=7
|
||
.fpu neon
|
||
|
||
.type mul_1x1_neon,%function
|
||
.align 5
|
||
mul_1x1_neon:
|
||
vshl.u64 `&Dlo("q1")`,d16,#8 @ q1-q3 are slided $a
|
||
vmull.p8 `&Q("d0")`,d16,d17 @ a<EFBFBD>bb
|
||
vshl.u64 `&Dlo("q2")`,d16,#16
|
||
vmull.p8 q1,`&Dlo("q1")`,d17 @ a<<8<EFBFBD>bb
|
||
vshl.u64 `&Dlo("q3")`,d16,#24
|
||
vmull.p8 q2,`&Dlo("q2")`,d17 @ a<<16<EFBFBD>bb
|
||
vshr.u64 `&Dlo("q1")`,#8
|
||
vmull.p8 q3,`&Dlo("q3")`,d17 @ a<<24<EFBFBD>bb
|
||
vshl.u64 `&Dhi("q1")`,#24
|
||
veor d0,`&Dlo("q1")`
|
||
vshr.u64 `&Dlo("q2")`,#16
|
||
veor d0,`&Dhi("q1")`
|
||
vshl.u64 `&Dhi("q2")`,#16
|
||
veor d0,`&Dlo("q2")`
|
||
vshr.u64 `&Dlo("q3")`,#24
|
||
veor d0,`&Dhi("q2")`
|
||
vshl.u64 `&Dhi("q3")`,#8
|
||
veor d0,`&Dlo("q3")`
|
||
veor d0,`&Dhi("q3")`
|
||
bx lr
|
||
.size mul_1x1_neon,.-mul_1x1_neon
|
||
#endif
|
||
___
|
||
################
|
||
# private interface to mul_1x1_ialu
|
||
#
|
||
$a="r1";
|
||
$b="r0";
|
||
|
||
($a0,$a1,$a2,$a12,$a4,$a14)=
|
||
($hi,$lo,$t0,$t1, $i0,$i1 )=map("r$_",(4..9),12);
|
||
|
||
$mask="r12";
|
||
|
||
$code.=<<___;
|
||
.type mul_1x1_ialu,%function
|
||
.align 5
|
||
mul_1x1_ialu:
|
||
mov $a0,#0
|
||
bic $a1,$a,#3<<30 @ a1=a&0x3fffffff
|
||
str $a0,[sp,#0] @ tab[0]=0
|
||
add $a2,$a1,$a1 @ a2=a1<<1
|
||
str $a1,[sp,#4] @ tab[1]=a1
|
||
eor $a12,$a1,$a2 @ a1^a2
|
||
str $a2,[sp,#8] @ tab[2]=a2
|
||
mov $a4,$a1,lsl#2 @ a4=a1<<2
|
||
str $a12,[sp,#12] @ tab[3]=a1^a2
|
||
eor $a14,$a1,$a4 @ a1^a4
|
||
str $a4,[sp,#16] @ tab[4]=a4
|
||
eor $a0,$a2,$a4 @ a2^a4
|
||
str $a14,[sp,#20] @ tab[5]=a1^a4
|
||
eor $a12,$a12,$a4 @ a1^a2^a4
|
||
str $a0,[sp,#24] @ tab[6]=a2^a4
|
||
and $i0,$mask,$b,lsl#2
|
||
str $a12,[sp,#28] @ tab[7]=a1^a2^a4
|
||
|
||
and $i1,$mask,$b,lsr#1
|
||
ldr $lo,[sp,$i0] @ tab[b & 0x7]
|
||
and $i0,$mask,$b,lsr#4
|
||
ldr $t1,[sp,$i1] @ tab[b >> 3 & 0x7]
|
||
and $i1,$mask,$b,lsr#7
|
||
ldr $t0,[sp,$i0] @ tab[b >> 6 & 0x7]
|
||
eor $lo,$lo,$t1,lsl#3 @ stall
|
||
mov $hi,$t1,lsr#29
|
||
ldr $t1,[sp,$i1] @ tab[b >> 9 & 0x7]
|
||
|
||
and $i0,$mask,$b,lsr#10
|
||
eor $lo,$lo,$t0,lsl#6
|
||
eor $hi,$hi,$t0,lsr#26
|
||
ldr $t0,[sp,$i0] @ tab[b >> 12 & 0x7]
|
||
|
||
and $i1,$mask,$b,lsr#13
|
||
eor $lo,$lo,$t1,lsl#9
|
||
eor $hi,$hi,$t1,lsr#23
|
||
ldr $t1,[sp,$i1] @ tab[b >> 15 & 0x7]
|
||
|
||
and $i0,$mask,$b,lsr#16
|
||
eor $lo,$lo,$t0,lsl#12
|
||
eor $hi,$hi,$t0,lsr#20
|
||
ldr $t0,[sp,$i0] @ tab[b >> 18 & 0x7]
|
||
|
||
and $i1,$mask,$b,lsr#19
|
||
eor $lo,$lo,$t1,lsl#15
|
||
eor $hi,$hi,$t1,lsr#17
|
||
ldr $t1,[sp,$i1] @ tab[b >> 21 & 0x7]
|
||
|
||
and $i0,$mask,$b,lsr#22
|
||
eor $lo,$lo,$t0,lsl#18
|
||
eor $hi,$hi,$t0,lsr#14
|
||
ldr $t0,[sp,$i0] @ tab[b >> 24 & 0x7]
|
||
|
||
and $i1,$mask,$b,lsr#25
|
||
eor $lo,$lo,$t1,lsl#21
|
||
eor $hi,$hi,$t1,lsr#11
|
||
ldr $t1,[sp,$i1] @ tab[b >> 27 & 0x7]
|
||
|
||
tst $a,#1<<30
|
||
and $i0,$mask,$b,lsr#28
|
||
eor $lo,$lo,$t0,lsl#24
|
||
eor $hi,$hi,$t0,lsr#8
|
||
ldr $t0,[sp,$i0] @ tab[b >> 30 ]
|
||
|
||
eorne $lo,$lo,$b,lsl#30
|
||
eorne $hi,$hi,$b,lsr#2
|
||
tst $a,#1<<31
|
||
eor $lo,$lo,$t1,lsl#27
|
||
eor $hi,$hi,$t1,lsr#5
|
||
eorne $lo,$lo,$b,lsl#31
|
||
eorne $hi,$hi,$b,lsr#1
|
||
eor $lo,$lo,$t0,lsl#30
|
||
eor $hi,$hi,$t0,lsr#2
|
||
|
||
mov pc,lr
|
||
.size mul_1x1_ialu,.-mul_1x1_ialu
|
||
___
|
||
################
|
||
# void bn_GF2m_mul_2x2(BN_ULONG *r,
|
||
# BN_ULONG a1,BN_ULONG a0,
|
||
# BN_ULONG b1,BN_ULONG b0); # r[3..0]=a1a0<EFBFBD>b1b0
|
||
|
||
($A1,$B1,$A0,$B0,$A1B1,$A0B0)=map("d$_",(18..23));
|
||
|
||
$code.=<<___;
|
||
.global bn_GF2m_mul_2x2
|
||
.type bn_GF2m_mul_2x2,%function
|
||
.align 5
|
||
bn_GF2m_mul_2x2:
|
||
#if __ARM_ARCH__>=7
|
||
ldr r12,.LOPENSSL_armcap
|
||
.Lpic: ldr r12,[pc,r12]
|
||
#ifdef __APPLE__
|
||
ldr r12,[r12]
|
||
#endif
|
||
tst r12,#1
|
||
beq .Lialu
|
||
|
||
veor $A1,$A1
|
||
#ifdef __APPLE__
|
||
vmov $B1,r3,r3 @ two copies of b1
|
||
#else
|
||
vmov.32 $B1,r3,r3 @ two copies of b1
|
||
#endif
|
||
vmov.32 ${A1}[0],r1 @ a1
|
||
|
||
veor $A0,$A0
|
||
vld1.32 ${B0}[],[sp,:32] @ two copies of b0
|
||
vmov.32 ${A0}[0],r2 @ a0
|
||
mov r12,lr
|
||
|
||
vmov d16,$A1
|
||
vmov d17,$B1
|
||
bl mul_1x1_neon @ a1<EFBFBD>b1
|
||
vmov $A1B1,d0
|
||
|
||
vmov d16,$A0
|
||
vmov d17,$B0
|
||
bl mul_1x1_neon @ a0<EFBFBD>b0
|
||
vmov $A0B0,d0
|
||
|
||
veor d16,$A0,$A1
|
||
veor d17,$B0,$B1
|
||
veor $A0,$A0B0,$A1B1
|
||
bl mul_1x1_neon @ (a0+a1)<EFBFBD>(b0+b1)
|
||
|
||
veor d0,$A0 @ (a0+a1)<EFBFBD>(b0+b1)-a0<EFBFBD>b0-a1<EFBFBD>b1
|
||
vshl.u64 d1,d0,#32
|
||
vshr.u64 d0,d0,#32
|
||
veor $A0B0,d1
|
||
veor $A1B1,d0
|
||
vst1.32 {${A0B0}[0]},[r0,:32]!
|
||
vst1.32 {${A0B0}[1]},[r0,:32]!
|
||
vst1.32 {${A1B1}[0]},[r0,:32]!
|
||
vst1.32 {${A1B1}[1]},[r0,:32]
|
||
bx r12
|
||
.align 4
|
||
.Lialu:
|
||
#endif
|
||
___
|
||
$ret="r10"; # reassigned 1st argument
|
||
$code.=<<___;
|
||
stmdb sp!,{r4-r10,lr}
|
||
mov $ret,r0 @ reassign 1st argument
|
||
mov $b,r3 @ $b=b1
|
||
ldr r3,[sp,#32] @ load b0
|
||
mov $mask,#7<<2
|
||
sub sp,sp,#32 @ allocate tab[8]
|
||
|
||
bl mul_1x1_ialu @ a1<EFBFBD>b1
|
||
str $lo,[$ret,#8]
|
||
str $hi,[$ret,#12]
|
||
|
||
eor $b,$b,r3 @ flip b0 and b1
|
||
eor $a,$a,r2 @ flip a0 and a1
|
||
eor r3,r3,$b
|
||
eor r2,r2,$a
|
||
eor $b,$b,r3
|
||
eor $a,$a,r2
|
||
bl mul_1x1_ialu @ a0<EFBFBD>b0
|
||
str $lo,[$ret]
|
||
str $hi,[$ret,#4]
|
||
|
||
eor $a,$a,r2
|
||
eor $b,$b,r3
|
||
bl mul_1x1_ialu @ (a1+a0)<EFBFBD>(b1+b0)
|
||
___
|
||
@r=map("r$_",(6..9));
|
||
$code.=<<___;
|
||
ldmia $ret,{@r[0]-@r[3]}
|
||
eor $lo,$lo,$hi
|
||
eor $hi,$hi,@r[1]
|
||
eor $lo,$lo,@r[0]
|
||
eor $hi,$hi,@r[2]
|
||
eor $lo,$lo,@r[3]
|
||
eor $hi,$hi,@r[3]
|
||
str $hi,[$ret,#8]
|
||
eor $lo,$lo,$hi
|
||
add sp,sp,#32 @ destroy tab[8]
|
||
str $lo,[$ret,#4]
|
||
|
||
#if __ARM_ARCH__>=5
|
||
ldmia sp!,{r4-r10,pc}
|
||
#else
|
||
ldmia sp!,{r4-r10,lr}
|
||
tst lr,#1
|
||
moveq pc,lr @ be binary compatible with V4, yet
|
||
bx lr @ interoperable with Thumb ISA:-)
|
||
#endif
|
||
.size bn_GF2m_mul_2x2,.-bn_GF2m_mul_2x2
|
||
#if __ARM_ARCH__>=7
|
||
.align 5
|
||
.LOPENSSL_armcap:
|
||
.word OPENSSL_armcap_P-(.Lpic+8)
|
||
#endif
|
||
.asciz "GF(2^m) Multiplication for ARMv4/NEON, CRYPTOGAMS by <appro\@openssl.org>"
|
||
.align 5
|
||
|
||
.comm OPENSSL_armcap_P,4,4
|
||
___
|
||
|
||
$code =~ s/\`([^\`]*)\`/eval $1/gem;
|
||
$code =~ s/\bbx\s+lr\b/.word\t0xe12fff1e/gm; # make it possible to compile with -march=armv4
|
||
print $code;
|
||
close STDOUT; # enforce flush
|