123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223 |
- #!/usr/bin/env perl
- # ====================================================================
- # Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL
- # project. The module is, however, dual licensed under OpenSSL and
- # CRYPTOGAMS licenses depending on where you obtain it. For further
- # details see http://www.openssl.org/~appro/cryptogams/.
- # ====================================================================
- # SHA256 block procedure for ARMv4. May 2007.
- # Performance is ~2x better than gcc 3.4 generated code and in "abso-
- # lute" terms is ~2250 cycles per 64-byte block or ~35 cycles per
- # byte [on single-issue Xscale PXA250 core].
- # July 2010.
- #
- # Rescheduling for dual-issue pipeline resulted in 22% improvement on
- # Cortex A8 core and ~20 cycles per processed byte.
- # February 2011.
- #
- # Profiler-assisted and platform-specific optimization resulted in 16%
- # improvement on Cortex A8 core and ~17 cycles per processed byte.
- $flavour = shift;
- if ($flavour=~/^\w[\w\-]*\.\w+$/) { $output=$flavour; undef $flavour; }
- else { while (($output=shift) && ($output!~/^\w[\w\-]*\.\w+$/)) {} }
- if ($flavour && $flavour ne "void") {
- $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
- ( $xlate="${dir}arm-xlate.pl" and -f $xlate ) or
- ( $xlate="${dir}../../perlasm/arm-xlate.pl" and -f $xlate) or
- die "can't locate arm-xlate.pl";
- open STDOUT,"| \"$^X\" $xlate $flavour $output";
- } else {
- open STDOUT,">$output";
- }
- $ctx="r0"; $t0="r0";
- $inp="r1"; $t3="r1";
- $len="r2"; $t1="r2";
- $T1="r3";
- $A="r4";
- $B="r5";
- $C="r6";
- $D="r7";
- $E="r8";
- $F="r9";
- $G="r10";
- $H="r11";
- @V=($A,$B,$C,$D,$E,$F,$G,$H);
- $t2="r12";
- $Ktbl="r14";
- @Sigma0=( 2,13,22);
- @Sigma1=( 6,11,25);
- @sigma0=( 7,18, 3);
- @sigma1=(17,19,10);
- sub BODY_00_15 {
- my ($i,$a,$b,$c,$d,$e,$f,$g,$h) = @_;
- $code.=<<___ if ($i<16);
- #if __ARM_ARCH__>=7
- ldr $T1,[$inp],#4
- #else
- ldrb $T1,[$inp,#3] @ $i
- ldrb $t2,[$inp,#2]
- ldrb $t1,[$inp,#1]
- ldrb $t0,[$inp],#4
- orr $T1,$T1,$t2,lsl#8
- orr $T1,$T1,$t1,lsl#16
- orr $T1,$T1,$t0,lsl#24
- #endif
- ___
- $code.=<<___;
- mov $t0,$e,ror#$Sigma1[0]
- ldr $t2,[$Ktbl],#4 @ *K256++
- eor $t0,$t0,$e,ror#$Sigma1[1]
- eor $t1,$f,$g
- #if $i>=16
- add $T1,$T1,$t3 @ from BODY_16_xx
- #elif __ARM_ARCH__>=7 && defined(__ARMEL__)
- rev $T1,$T1
- #endif
- #if $i==15
- str $inp,[sp,#17*4] @ leave room for $t3
- #endif
- eor $t0,$t0,$e,ror#$Sigma1[2] @ Sigma1(e)
- and $t1,$t1,$e
- str $T1,[sp,#`$i%16`*4]
- add $T1,$T1,$t0
- eor $t1,$t1,$g @ Ch(e,f,g)
- add $T1,$T1,$h
- mov $h,$a,ror#$Sigma0[0]
- add $T1,$T1,$t1
- eor $h,$h,$a,ror#$Sigma0[1]
- add $T1,$T1,$t2
- eor $h,$h,$a,ror#$Sigma0[2] @ Sigma0(a)
- #if $i>=15
- ldr $t3,[sp,#`($i+2)%16`*4] @ from BODY_16_xx
- #endif
- orr $t0,$a,$b
- and $t1,$a,$b
- and $t0,$t0,$c
- add $h,$h,$T1
- orr $t0,$t0,$t1 @ Maj(a,b,c)
- add $d,$d,$T1
- add $h,$h,$t0
- ___
- }
- sub BODY_16_XX {
- my ($i,$a,$b,$c,$d,$e,$f,$g,$h) = @_;
- $code.=<<___;
- @ ldr $t3,[sp,#`($i+1)%16`*4] @ $i
- ldr $t2,[sp,#`($i+14)%16`*4]
- mov $t0,$t3,ror#$sigma0[0]
- ldr $T1,[sp,#`($i+0)%16`*4]
- eor $t0,$t0,$t3,ror#$sigma0[1]
- ldr $t1,[sp,#`($i+9)%16`*4]
- eor $t0,$t0,$t3,lsr#$sigma0[2] @ sigma0(X[i+1])
- mov $t3,$t2,ror#$sigma1[0]
- add $T1,$T1,$t0
- eor $t3,$t3,$t2,ror#$sigma1[1]
- add $T1,$T1,$t1
- eor $t3,$t3,$t2,lsr#$sigma1[2] @ sigma1(X[i+14])
- @ add $T1,$T1,$t3
- ___
- &BODY_00_15(@_);
- }
- $code=<<___;
- #include "arm_arch.h"
- .text
- .code 32
- .type K256,%object
- .align 5
- K256:
- .word 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
- .word 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
- .word 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
- .word 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
- .word 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
- .word 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
- .word 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
- .word 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
- .word 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
- .word 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
- .word 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
- .word 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
- .word 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
- .word 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
- .word 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
- .word 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
- .size K256,.-K256
- .global sha256_block_data_order
- .type sha256_block_data_order,%function
- sha256_block_data_order:
- sub r3,pc,#8 @ sha256_block_data_order
- add $len,$inp,$len,lsl#6 @ len to point at the end of inp
- stmdb sp!,{$ctx,$inp,$len,r4-r11,lr}
- ldmia $ctx,{$A,$B,$C,$D,$E,$F,$G,$H}
- sub $Ktbl,r3,#256 @ K256
- sub sp,sp,#16*4 @ alloca(X[16])
- .Loop:
- ___
- for($i=0;$i<16;$i++) { &BODY_00_15($i,@V); unshift(@V,pop(@V)); }
- $code.=".Lrounds_16_xx:\n";
- for (;$i<32;$i++) { &BODY_16_XX($i,@V); unshift(@V,pop(@V)); }
- $code.=<<___;
- and $t2,$t2,#0xff
- cmp $t2,#0xf2
- bne .Lrounds_16_xx
- ldr $T1,[sp,#16*4] @ pull ctx
- ldr $t0,[$T1,#0]
- ldr $t1,[$T1,#4]
- ldr $t2,[$T1,#8]
- add $A,$A,$t0
- ldr $t0,[$T1,#12]
- add $B,$B,$t1
- ldr $t1,[$T1,#16]
- add $C,$C,$t2
- ldr $t2,[$T1,#20]
- add $D,$D,$t0
- ldr $t0,[$T1,#24]
- add $E,$E,$t1
- ldr $t1,[$T1,#28]
- add $F,$F,$t2
- ldr $inp,[sp,#17*4] @ pull inp
- ldr $t2,[sp,#18*4] @ pull inp+len
- add $G,$G,$t0
- add $H,$H,$t1
- stmia $T1,{$A,$B,$C,$D,$E,$F,$G,$H}
- cmp $inp,$t2
- sub $Ktbl,$Ktbl,#256 @ rewind Ktbl
- bne .Loop
- add sp,sp,#`16+3`*4 @ destroy frame
- #if __ARM_ARCH__>=5
- ldmia sp!,{r4-r11,pc}
- #else
- ldmia sp!,{r4-r11,lr}
- tst lr,#1
- moveq pc,lr @ be binary compatible with V4, yet
- bx lr @ interoperable with Thumb ISA:-)
- #endif
- .size sha256_block_data_order,.-sha256_block_data_order
- .asciz "SHA256 block transform for ARMv4, CRYPTOGAMS by <appro\@openssl.org>"
- .align 2
- ___
- $code =~ s/\`([^\`]*)\`/eval $1/gem;
- $code =~ s/\bbx\s+lr\b/.word\t0xe12fff1e/gm; # make it possible to compile with -march=armv4
- print $code;
- close STDOUT; # enforce flush
|