123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611 |
- #!/usr/bin/env perl
- # Copyright 2017-2020 The OpenSSL Project Authors. All Rights Reserved.
- #
- # Licensed under the Apache License 2.0 (the "License"). You may not use
- # this file except in compliance with the License. You can obtain a copy
- # in the file LICENSE in the source distribution or at
- # https://www.openssl.org/source/license.html
- #
- # ====================================================================
- # Written by Andy Polyakov <appro@openssl.org> for the OpenSSL
- # project. The module is, however, dual licensed under OpenSSL and
- # CRYPTOGAMS licenses depending on where you obtain it. For further
- # details see http://www.openssl.org/~appro/cryptogams/.
- # ====================================================================
- #
- # Keccak-1600 for x86_64.
- #
- # June 2017.
- #
- # Below code is [lane complementing] KECCAK_2X implementation (see
- # sha/keccak1600.c) with C[5] and D[5] held in register bank. Though
- # instead of actually unrolling the loop pair-wise I simply flip
- # pointers to T[][] and A[][] at the end of round. Since number of
- # rounds is even, last round writes to A[][] and everything works out.
- # How does it compare to x86_64 assembly module in Keccak Code Package?
- # Depending on processor it's either as fast or faster by up to 15%...
- #
- ########################################################################
- # Numbers are cycles per processed byte out of large message.
- #
- # r=1088(*)
- #
- # P4 25.8
- # Core 2 12.9
- # Westmere 13.7
- # Sandy Bridge 12.9(**)
- # Haswell 9.6
- # Skylake 9.4
- # Silvermont 22.8
- # Goldmont 15.8
- # VIA Nano 17.3
- # Sledgehammer 13.3
- # Bulldozer 16.5
- # Ryzen 8.8
- #
- # (*) Corresponds to SHA3-256. Improvement over compiler-generate
- # varies a lot, most common coefficient is 15% in comparison to
- # gcc-5.x, 50% for gcc-4.x, 90% for gcc-3.x.
- # (**) Sandy Bridge has broken rotate instruction. Performance can be
- # improved by 14% by replacing rotates with double-precision
- # shift with same register as source and destination.
- # $output is the last argument if it looks like a file (it has an extension)
- # $flavour is the first argument if it doesn't look like a file
- $output = $#ARGV >= 0 && $ARGV[$#ARGV] =~ m|\.\w+$| ? pop : undef;
- $flavour = $#ARGV >= 0 && $ARGV[0] !~ m|\.| ? shift : undef;
- $win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/);
- $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
- ( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or
- ( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or
- die "can't locate x86_64-xlate.pl";
- open OUT,"| \"$^X\" \"$xlate\" $flavour \"$output\""
- or die "can't call $xlate: $!";
- *STDOUT=*OUT;
- my @A = map([ 8*$_-100, 8*($_+1)-100, 8*($_+2)-100,
- 8*($_+3)-100, 8*($_+4)-100 ], (0,5,10,15,20));
- my @C = ("%rax","%rbx","%rcx","%rdx","%rbp");
- my @D = map("%r$_",(8..12));
- my @T = map("%r$_",(13..14));
- my $iotas = "%r15";
- my @rhotates = ([ 0, 1, 62, 28, 27 ],
- [ 36, 44, 6, 55, 20 ],
- [ 3, 10, 43, 25, 39 ],
- [ 41, 45, 15, 21, 8 ],
- [ 18, 2, 61, 56, 14 ]);
- $code.=<<___;
- .text
- .type __KeccakF1600,\@abi-omnipotent
- .align 32
- __KeccakF1600:
- .cfi_startproc
- mov $A[4][0](%rdi),@C[0]
- mov $A[4][1](%rdi),@C[1]
- mov $A[4][2](%rdi),@C[2]
- mov $A[4][3](%rdi),@C[3]
- mov $A[4][4](%rdi),@C[4]
- jmp .Loop
- .align 32
- .Loop:
- mov $A[0][0](%rdi),@D[0]
- mov $A[1][1](%rdi),@D[1]
- mov $A[2][2](%rdi),@D[2]
- mov $A[3][3](%rdi),@D[3]
- xor $A[0][2](%rdi),@C[2]
- xor $A[0][3](%rdi),@C[3]
- xor @D[0], @C[0]
- xor $A[0][1](%rdi),@C[1]
- xor $A[1][2](%rdi),@C[2]
- xor $A[1][0](%rdi),@C[0]
- mov @C[4],@D[4]
- xor $A[0][4](%rdi),@C[4]
- xor @D[2], @C[2]
- xor $A[2][0](%rdi),@C[0]
- xor $A[1][3](%rdi),@C[3]
- xor @D[1], @C[1]
- xor $A[1][4](%rdi),@C[4]
- xor $A[3][2](%rdi),@C[2]
- xor $A[3][0](%rdi),@C[0]
- xor $A[2][3](%rdi),@C[3]
- xor $A[2][1](%rdi),@C[1]
- xor $A[2][4](%rdi),@C[4]
- mov @C[2],@T[0]
- rol \$1,@C[2]
- xor @C[0],@C[2] # D[1] = ROL64(C[2], 1) ^ C[0]
- xor @D[3], @C[3]
- rol \$1,@C[0]
- xor @C[3],@C[0] # D[4] = ROL64(C[0], 1) ^ C[3]
- xor $A[3][1](%rdi),@C[1]
- rol \$1,@C[3]
- xor @C[1],@C[3] # D[2] = ROL64(C[3], 1) ^ C[1]
- xor $A[3][4](%rdi),@C[4]
- rol \$1,@C[1]
- xor @C[4],@C[1] # D[0] = ROL64(C[1], 1) ^ C[4]
- rol \$1,@C[4]
- xor @T[0],@C[4] # D[3] = ROL64(C[4], 1) ^ C[2]
- ___
- (@D[0..4], @C) = (@C[1..4,0], @D);
- $code.=<<___;
- xor @D[1],@C[1]
- xor @D[2],@C[2]
- rol \$$rhotates[1][1],@C[1]
- xor @D[3],@C[3]
- xor @D[4],@C[4]
- rol \$$rhotates[2][2],@C[2]
- xor @D[0],@C[0]
- mov @C[1],@T[0]
- rol \$$rhotates[3][3],@C[3]
- or @C[2],@C[1]
- xor @C[0],@C[1] # C[0] ^ ( C[1] | C[2])
- rol \$$rhotates[4][4],@C[4]
- xor ($iotas),@C[1]
- lea 8($iotas),$iotas
- mov @C[4],@T[1]
- and @C[3],@C[4]
- mov @C[1],$A[0][0](%rsi) # R[0][0] = C[0] ^ ( C[1] | C[2]) ^ iotas[i]
- xor @C[2],@C[4] # C[2] ^ ( C[4] & C[3])
- not @C[2]
- mov @C[4],$A[0][2](%rsi) # R[0][2] = C[2] ^ ( C[4] & C[3])
- or @C[3],@C[2]
- mov $A[4][2](%rdi),@C[4]
- xor @T[0],@C[2] # C[1] ^ (~C[2] | C[3])
- mov @C[2],$A[0][1](%rsi) # R[0][1] = C[1] ^ (~C[2] | C[3])
- and @C[0],@T[0]
- mov $A[1][4](%rdi),@C[1]
- xor @T[1],@T[0] # C[4] ^ ( C[1] & C[0])
- mov $A[2][0](%rdi),@C[2]
- mov @T[0],$A[0][4](%rsi) # R[0][4] = C[4] ^ ( C[1] & C[0])
- or @C[0],@T[1]
- mov $A[0][3](%rdi),@C[0]
- xor @C[3],@T[1] # C[3] ^ ( C[4] | C[0])
- mov $A[3][1](%rdi),@C[3]
- mov @T[1],$A[0][3](%rsi) # R[0][3] = C[3] ^ ( C[4] | C[0])
- xor @D[3],@C[0]
- xor @D[2],@C[4]
- rol \$$rhotates[0][3],@C[0]
- xor @D[1],@C[3]
- xor @D[4],@C[1]
- rol \$$rhotates[4][2],@C[4]
- rol \$$rhotates[3][1],@C[3]
- xor @D[0],@C[2]
- rol \$$rhotates[1][4],@C[1]
- mov @C[0],@T[0]
- or @C[4],@C[0]
- rol \$$rhotates[2][0],@C[2]
- xor @C[3],@C[0] # C[3] ^ (C[0] | C[4])
- mov @C[0],$A[1][3](%rsi) # R[1][3] = C[3] ^ (C[0] | C[4])
- mov @C[1],@T[1]
- and @T[0],@C[1]
- mov $A[0][1](%rdi),@C[0]
- xor @C[4],@C[1] # C[4] ^ (C[1] & C[0])
- not @C[4]
- mov @C[1],$A[1][4](%rsi) # R[1][4] = C[4] ^ (C[1] & C[0])
- or @C[3],@C[4]
- mov $A[1][2](%rdi),@C[1]
- xor @C[2],@C[4] # C[2] ^ (~C[4] | C[3])
- mov @C[4],$A[1][2](%rsi) # R[1][2] = C[2] ^ (~C[4] | C[3])
- and @C[2],@C[3]
- mov $A[4][0](%rdi),@C[4]
- xor @T[1],@C[3] # C[1] ^ (C[3] & C[2])
- mov @C[3],$A[1][1](%rsi) # R[1][1] = C[1] ^ (C[3] & C[2])
- or @C[2],@T[1]
- mov $A[2][3](%rdi),@C[2]
- xor @T[0],@T[1] # C[0] ^ (C[1] | C[2])
- mov $A[3][4](%rdi),@C[3]
- mov @T[1],$A[1][0](%rsi) # R[1][0] = C[0] ^ (C[1] | C[2])
- xor @D[3],@C[2]
- xor @D[4],@C[3]
- rol \$$rhotates[2][3],@C[2]
- xor @D[2],@C[1]
- rol \$$rhotates[3][4],@C[3]
- xor @D[0],@C[4]
- rol \$$rhotates[1][2],@C[1]
- xor @D[1],@C[0]
- rol \$$rhotates[4][0],@C[4]
- mov @C[2],@T[0]
- and @C[3],@C[2]
- rol \$$rhotates[0][1],@C[0]
- not @C[3]
- xor @C[1],@C[2] # C[1] ^ ( C[2] & C[3])
- mov @C[2],$A[2][1](%rsi) # R[2][1] = C[1] ^ ( C[2] & C[3])
- mov @C[4],@T[1]
- and @C[3],@C[4]
- mov $A[2][1](%rdi),@C[2]
- xor @T[0],@C[4] # C[2] ^ ( C[4] & ~C[3])
- mov @C[4],$A[2][2](%rsi) # R[2][2] = C[2] ^ ( C[4] & ~C[3])
- or @C[1],@T[0]
- mov $A[4][3](%rdi),@C[4]
- xor @C[0],@T[0] # C[0] ^ ( C[2] | C[1])
- mov @T[0],$A[2][0](%rsi) # R[2][0] = C[0] ^ ( C[2] | C[1])
- and @C[0],@C[1]
- xor @T[1],@C[1] # C[4] ^ ( C[1] & C[0])
- mov @C[1],$A[2][4](%rsi) # R[2][4] = C[4] ^ ( C[1] & C[0])
- or @C[0],@T[1]
- mov $A[1][0](%rdi),@C[1]
- xor @C[3],@T[1] # ~C[3] ^ ( C[0] | C[4])
- mov $A[3][2](%rdi),@C[3]
- mov @T[1],$A[2][3](%rsi) # R[2][3] = ~C[3] ^ ( C[0] | C[4])
- mov $A[0][4](%rdi),@C[0]
- xor @D[1],@C[2]
- xor @D[2],@C[3]
- rol \$$rhotates[2][1],@C[2]
- xor @D[0],@C[1]
- rol \$$rhotates[3][2],@C[3]
- xor @D[3],@C[4]
- rol \$$rhotates[1][0],@C[1]
- xor @D[4],@C[0]
- rol \$$rhotates[4][3],@C[4]
- mov @C[2],@T[0]
- or @C[3],@C[2]
- rol \$$rhotates[0][4],@C[0]
- not @C[3]
- xor @C[1],@C[2] # C[1] ^ ( C[2] | C[3])
- mov @C[2],$A[3][1](%rsi) # R[3][1] = C[1] ^ ( C[2] | C[3])
- mov @C[4],@T[1]
- or @C[3],@C[4]
- xor @T[0],@C[4] # C[2] ^ ( C[4] | ~C[3])
- mov @C[4],$A[3][2](%rsi) # R[3][2] = C[2] ^ ( C[4] | ~C[3])
- and @C[1],@T[0]
- xor @C[0],@T[0] # C[0] ^ ( C[2] & C[1])
- mov @T[0],$A[3][0](%rsi) # R[3][0] = C[0] ^ ( C[2] & C[1])
- or @C[0],@C[1]
- xor @T[1],@C[1] # C[4] ^ ( C[1] | C[0])
- mov @C[1],$A[3][4](%rsi) # R[3][4] = C[4] ^ ( C[1] | C[0])
- and @T[1],@C[0]
- xor @C[3],@C[0] # ~C[3] ^ ( C[0] & C[4])
- mov @C[0],$A[3][3](%rsi) # R[3][3] = ~C[3] ^ ( C[0] & C[4])
- xor $A[0][2](%rdi),@D[2]
- xor $A[1][3](%rdi),@D[3]
- rol \$$rhotates[0][2],@D[2]
- xor $A[4][1](%rdi),@D[1]
- rol \$$rhotates[1][3],@D[3]
- xor $A[2][4](%rdi),@D[4]
- rol \$$rhotates[4][1],@D[1]
- xor $A[3][0](%rdi),@D[0]
- xchg %rsi,%rdi
- rol \$$rhotates[2][4],@D[4]
- rol \$$rhotates[3][0],@D[0]
- ___
- @C = @D[2..4,0,1];
- $code.=<<___;
- mov @C[0],@T[0]
- and @C[1],@C[0]
- not @C[1]
- xor @C[4],@C[0] # C[4] ^ ( C[0] & C[1])
- mov @C[0],$A[4][4](%rdi) # R[4][4] = C[4] ^ ( C[0] & C[1])
- mov @C[2],@T[1]
- and @C[1],@C[2]
- xor @T[0],@C[2] # C[0] ^ ( C[2] & ~C[1])
- mov @C[2],$A[4][0](%rdi) # R[4][0] = C[0] ^ ( C[2] & ~C[1])
- or @C[4],@T[0]
- xor @C[3],@T[0] # C[3] ^ ( C[0] | C[4])
- mov @T[0],$A[4][3](%rdi) # R[4][3] = C[3] ^ ( C[0] | C[4])
- and @C[3],@C[4]
- xor @T[1],@C[4] # C[2] ^ ( C[4] & C[3])
- mov @C[4],$A[4][2](%rdi) # R[4][2] = C[2] ^ ( C[4] & C[3])
- or @T[1],@C[3]
- xor @C[1],@C[3] # ~C[1] ^ ( C[2] | C[3])
- mov @C[3],$A[4][1](%rdi) # R[4][1] = ~C[1] ^ ( C[2] | C[3])
- mov @C[0],@C[1] # harmonize with the loop top
- mov @T[0],@C[0]
- test \$255,$iotas
- jnz .Loop
- lea -192($iotas),$iotas # rewind iotas
- ret
- .cfi_endproc
- .size __KeccakF1600,.-__KeccakF1600
- .type KeccakF1600,\@abi-omnipotent
- .align 32
- KeccakF1600:
- .cfi_startproc
- push %rbx
- .cfi_push %rbx
- push %rbp
- .cfi_push %rbp
- push %r12
- .cfi_push %r12
- push %r13
- .cfi_push %r13
- push %r14
- .cfi_push %r14
- push %r15
- .cfi_push %r15
- lea 100(%rdi),%rdi # size optimization
- sub \$200,%rsp
- .cfi_adjust_cfa_offset 200
- notq $A[0][1](%rdi)
- notq $A[0][2](%rdi)
- notq $A[1][3](%rdi)
- notq $A[2][2](%rdi)
- notq $A[3][2](%rdi)
- notq $A[4][0](%rdi)
- lea iotas(%rip),$iotas
- lea 100(%rsp),%rsi # size optimization
- call __KeccakF1600
- notq $A[0][1](%rdi)
- notq $A[0][2](%rdi)
- notq $A[1][3](%rdi)
- notq $A[2][2](%rdi)
- notq $A[3][2](%rdi)
- notq $A[4][0](%rdi)
- lea -100(%rdi),%rdi # preserve A[][]
- add \$200,%rsp
- .cfi_adjust_cfa_offset -200
- pop %r15
- .cfi_pop %r15
- pop %r14
- .cfi_pop %r14
- pop %r13
- .cfi_pop %r13
- pop %r12
- .cfi_pop %r12
- pop %rbp
- .cfi_pop %rbp
- pop %rbx
- .cfi_pop %rbx
- ret
- .cfi_endproc
- .size KeccakF1600,.-KeccakF1600
- ___
- { my ($A_flat,$inp,$len,$bsz) = ("%rdi","%rsi","%rdx","%rcx");
- ($A_flat,$inp) = ("%r8","%r9");
- $code.=<<___;
- .globl SHA3_absorb
- .type SHA3_absorb,\@function,4
- .align 32
- SHA3_absorb:
- .cfi_startproc
- push %rbx
- .cfi_push %rbx
- push %rbp
- .cfi_push %rbp
- push %r12
- .cfi_push %r12
- push %r13
- .cfi_push %r13
- push %r14
- .cfi_push %r14
- push %r15
- .cfi_push %r15
- lea 100(%rdi),%rdi # size optimization
- sub \$232,%rsp
- .cfi_adjust_cfa_offset 232
- mov %rsi,$inp
- lea 100(%rsp),%rsi # size optimization
- notq $A[0][1](%rdi)
- notq $A[0][2](%rdi)
- notq $A[1][3](%rdi)
- notq $A[2][2](%rdi)
- notq $A[3][2](%rdi)
- notq $A[4][0](%rdi)
- lea iotas(%rip),$iotas
- mov $bsz,216-100(%rsi) # save bsz
- .Loop_absorb:
- cmp $bsz,$len
- jc .Ldone_absorb
- shr \$3,$bsz
- lea -100(%rdi),$A_flat
- .Lblock_absorb:
- mov ($inp),%rax
- lea 8($inp),$inp
- xor ($A_flat),%rax
- lea 8($A_flat),$A_flat
- sub \$8,$len
- mov %rax,-8($A_flat)
- sub \$1,$bsz
- jnz .Lblock_absorb
- mov $inp,200-100(%rsi) # save inp
- mov $len,208-100(%rsi) # save len
- call __KeccakF1600
- mov 200-100(%rsi),$inp # pull inp
- mov 208-100(%rsi),$len # pull len
- mov 216-100(%rsi),$bsz # pull bsz
- jmp .Loop_absorb
- .align 32
- .Ldone_absorb:
- mov $len,%rax # return value
- notq $A[0][1](%rdi)
- notq $A[0][2](%rdi)
- notq $A[1][3](%rdi)
- notq $A[2][2](%rdi)
- notq $A[3][2](%rdi)
- notq $A[4][0](%rdi)
- add \$232,%rsp
- .cfi_adjust_cfa_offset -232
- pop %r15
- .cfi_pop %r15
- pop %r14
- .cfi_pop %r14
- pop %r13
- .cfi_pop %r13
- pop %r12
- .cfi_pop %r12
- pop %rbp
- .cfi_pop %rbp
- pop %rbx
- .cfi_pop %rbx
- ret
- .cfi_endproc
- .size SHA3_absorb,.-SHA3_absorb
- ___
- }
- { my ($A_flat,$out,$len,$bsz) = ("%rdi","%rsi","%rdx","%rcx");
- ($out,$len,$bsz) = ("%r12","%r13","%r14");
- $code.=<<___;
- .globl SHA3_squeeze
- .type SHA3_squeeze,\@function,4
- .align 32
- SHA3_squeeze:
- .cfi_startproc
- push %r12
- .cfi_push %r12
- push %r13
- .cfi_push %r13
- push %r14
- .cfi_push %r14
- shr \$3,%rcx
- mov $A_flat,%r8
- mov %rsi,$out
- mov %rdx,$len
- mov %rcx,$bsz
- jmp .Loop_squeeze
- .align 32
- .Loop_squeeze:
- cmp \$8,$len
- jb .Ltail_squeeze
- mov (%r8),%rax
- lea 8(%r8),%r8
- mov %rax,($out)
- lea 8($out),$out
- sub \$8,$len # len -= 8
- jz .Ldone_squeeze
- sub \$1,%rcx # bsz--
- jnz .Loop_squeeze
- call KeccakF1600
- mov $A_flat,%r8
- mov $bsz,%rcx
- jmp .Loop_squeeze
- .Ltail_squeeze:
- mov %r8, %rsi
- mov $out,%rdi
- mov $len,%rcx
- .byte 0xf3,0xa4 # rep movsb
- .Ldone_squeeze:
- pop %r14
- .cfi_pop %r14
- pop %r13
- .cfi_pop %r13
- pop %r12
- .cfi_pop %r13
- ret
- .cfi_endproc
- .size SHA3_squeeze,.-SHA3_squeeze
- ___
- }
- $code.=<<___;
- .align 256
- .quad 0,0,0,0,0,0,0,0
- .type iotas,\@object
- iotas:
- .quad 0x0000000000000001
- .quad 0x0000000000008082
- .quad 0x800000000000808a
- .quad 0x8000000080008000
- .quad 0x000000000000808b
- .quad 0x0000000080000001
- .quad 0x8000000080008081
- .quad 0x8000000000008009
- .quad 0x000000000000008a
- .quad 0x0000000000000088
- .quad 0x0000000080008009
- .quad 0x000000008000000a
- .quad 0x000000008000808b
- .quad 0x800000000000008b
- .quad 0x8000000000008089
- .quad 0x8000000000008003
- .quad 0x8000000000008002
- .quad 0x8000000000000080
- .quad 0x000000000000800a
- .quad 0x800000008000000a
- .quad 0x8000000080008081
- .quad 0x8000000000008080
- .quad 0x0000000080000001
- .quad 0x8000000080008008
- .size iotas,.-iotas
- .asciz "Keccak-1600 absorb and squeeze for x86_64, CRYPTOGAMS by <appro\@openssl.org>"
- ___
- foreach (split("\n",$code)) {
- # Below replacement results in 11.2 on Sandy Bridge, 9.4 on
- # Haswell, but it hurts other processors by up to 2-3-4x...
- #s/rol\s+(\$[0-9]+),(%[a-z][a-z0-9]+)/shld\t$1,$2,$2/;
- # Below replacement results in 9.3 on Haswell [as well as
- # on Ryzen, i.e. it *hurts* Ryzen]...
- #s/rol\s+\$([0-9]+),(%[a-z][a-z0-9]+)/rorx\t\$64-$1,$2,$2/;
- print $_, "\n";
- }
- close STDOUT or die "error closing STDOUT: $!";
|