#!/usr/bin/env perl # # ==================================================================== # Written by Andy Polyakov for the OpenSSL # project. The module is, however, dual licensed under OpenSSL and # CRYPTOGAMS licenses depending on where you obtain it. For further # details see http://www.openssl.org/~appro/cryptogams/. # ==================================================================== # # Keccak-1600 for x86_64. # # June 2017. # # Below code is [lane complementing] KECCAK_2X implementation (see # sha/keccak1600.c) with C[5] and D[5] held in register bank. Though # instead of actually unrolling the loop pair-wise I simply flip # pointers to T[][] and A[][] at the end of round. Since number of # rounds is even, last round writes to A[][] and everything works out. # How does it compare to x86_64 assembly module in Keccak Code Package? # Depending on processor it's either as fast or faster by up to 15%... # ######################################################################## # Numbers are cycles per processed byte out of large message. # # r=1088(*) # # P4 25.8 # Core 2 12.9 # Westmere 13.7 # Sandy Bridge 12.9(**) # Haswell 9.6 # Skylake 9.4 # Ice Lake 8.6 # Silvermont 22.8 # Goldmont 15.8 # VIA Nano 17.3 # Sledgehammer 13.3 # Bulldozer 16.5 # Ryzen 8.8 # Zen 4 7.6 # # (*) Corresponds to SHA3-256. Improvement over compiler-generate # varies a lot, most commont coefficient is 15% in comparison to # gcc-5.x, 50% for gcc-4.x, 90% for gcc-3.x. # (**) Sandy Bridge has broken rotate instruction. Performance can be # improved by 14% by replacing rotates with double-precision # shift with same register as source and destination. $flavour = shift; $output = shift; if ($flavour =~ /\./) { $output = $flavour; undef $flavour; } $win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/); $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1; ( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or ( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or die "can't locate x86_64-xlate.pl"; open OUT,"| \"$^X\" \"$xlate\" $flavour \"$output\""; *STDOUT=*OUT; my @A = map([ 8*$_-100, 8*($_+1)-100, 8*($_+2)-100, 8*($_+3)-100, 8*($_+4)-100 ], (0,5,10,15,20)); my @C = ("%rax","%rbx","%rcx","%rdx","%rbp"); my @D = map("%r$_",(8..12)); my @T = map("%r$_",(13..14)); my $iotas = "%r15"; my @rhotates = ([ 0, 1, 62, 28, 27 ], [ 36, 44, 6, 55, 20 ], [ 3, 10, 43, 25, 39 ], [ 41, 45, 15, 21, 8 ], [ 18, 2, 61, 56, 14 ]); $code.=<<___; .text .type __KeccakF1600,\@abi-omnipotent .align 32 __KeccakF1600: mov $A[4][0](%rdi),@C[0] mov $A[4][1](%rdi),@C[1] mov $A[4][2](%rdi),@C[2] mov $A[4][3](%rdi),@C[3] mov $A[4][4](%rdi),@C[4] jmp .Loop .align 32 .Loop: mov $A[0][0](%rdi),@D[0] mov $A[1][1](%rdi),@D[1] mov $A[2][2](%rdi),@D[2] mov $A[3][3](%rdi),@D[3] xor $A[0][2](%rdi),@C[2] xor $A[0][3](%rdi),@C[3] xor @D[0], @C[0] xor $A[0][1](%rdi),@C[1] xor $A[1][2](%rdi),@C[2] xor $A[1][0](%rdi),@C[0] mov @C[4],@D[4] xor $A[0][4](%rdi),@C[4] xor @D[2], @C[2] xor $A[2][0](%rdi),@C[0] xor $A[1][3](%rdi),@C[3] xor @D[1], @C[1] xor $A[1][4](%rdi),@C[4] xor $A[3][2](%rdi),@C[2] xor $A[3][0](%rdi),@C[0] xor $A[2][3](%rdi),@C[3] xor $A[2][1](%rdi),@C[1] xor $A[2][4](%rdi),@C[4] mov @C[2],@T[0] rol \$1,@C[2] xor @C[0],@C[2] # D[1] = ROL64(C[2], 1) ^ C[0] xor @D[3], @C[3] rol \$1,@C[0] xor @C[3],@C[0] # D[4] = ROL64(C[0], 1) ^ C[3] xor $A[3][1](%rdi),@C[1] rol \$1,@C[3] xor @C[1],@C[3] # D[2] = ROL64(C[3], 1) ^ C[1] xor $A[3][4](%rdi),@C[4] rol \$1,@C[1] xor @C[4],@C[1] # D[0] = ROL64(C[1], 1) ^ C[4] rol \$1,@C[4] xor @T[0],@C[4] # D[3] = ROL64(C[4], 1) ^ C[2] ___ (@D[0..4], @C) = (@C[1..4,0], @D); $code.=<<___; xor @D[1],@C[1] xor @D[2],@C[2] rol \$$rhotates[1][1],@C[1] xor @D[3],@C[3] xor @D[4],@C[4] rol \$$rhotates[2][2],@C[2] xor @D[0],@C[0] mov @C[1],@T[0] rol \$$rhotates[3][3],@C[3] or @C[2],@C[1] xor @C[0],@C[1] # C[0] ^ ( C[1] | C[2]) rol \$$rhotates[4][4],@C[4] xor ($iotas),@C[1] lea 8($iotas),$iotas mov @C[4],@T[1] and @C[3],@C[4] mov @C[1],$A[0][0](%rsi) # R[0][0] = C[0] ^ ( C[1] | C[2]) ^ iotas[i] xor @C[2],@C[4] # C[2] ^ ( C[4] & C[3]) not @C[2] mov @C[4],$A[0][2](%rsi) # R[0][2] = C[2] ^ ( C[4] & C[3]) or @C[3],@C[2] mov $A[4][2](%rdi),@C[4] xor @T[0],@C[2] # C[1] ^ (~C[2] | C[3]) mov @C[2],$A[0][1](%rsi) # R[0][1] = C[1] ^ (~C[2] | C[3]) and @C[0],@T[0] mov $A[1][4](%rdi),@C[1] xor @T[1],@T[0] # C[4] ^ ( C[1] & C[0]) mov $A[2][0](%rdi),@C[2] mov @T[0],$A[0][4](%rsi) # R[0][4] = C[4] ^ ( C[1] & C[0]) or @C[0],@T[1] mov $A[0][3](%rdi),@C[0] xor @C[3],@T[1] # C[3] ^ ( C[4] | C[0]) mov $A[3][1](%rdi),@C[3] mov @T[1],$A[0][3](%rsi) # R[0][3] = C[3] ^ ( C[4] | C[0]) xor @D[3],@C[0] xor @D[2],@C[4] rol \$$rhotates[0][3],@C[0] xor @D[1],@C[3] xor @D[4],@C[1] rol \$$rhotates[4][2],@C[4] rol \$$rhotates[3][1],@C[3] xor @D[0],@C[2] rol \$$rhotates[1][4],@C[1] mov @C[0],@T[0] or @C[4],@C[0] rol \$$rhotates[2][0],@C[2] xor @C[3],@C[0] # C[3] ^ (C[0] | C[4]) mov @C[0],$A[1][3](%rsi) # R[1][3] = C[3] ^ (C[0] | C[4]) mov @C[1],@T[1] and @T[0],@C[1] mov $A[0][1](%rdi),@C[0] xor @C[4],@C[1] # C[4] ^ (C[1] & C[0]) not @C[4] mov @C[1],$A[1][4](%rsi) # R[1][4] = C[4] ^ (C[1] & C[0]) or @C[3],@C[4] mov $A[1][2](%rdi),@C[1] xor @C[2],@C[4] # C[2] ^ (~C[4] | C[3]) mov @C[4],$A[1][2](%rsi) # R[1][2] = C[2] ^ (~C[4] | C[3]) and @C[2],@C[3] mov $A[4][0](%rdi),@C[4] xor @T[1],@C[3] # C[1] ^ (C[3] & C[2]) mov @C[3],$A[1][1](%rsi) # R[1][1] = C[1] ^ (C[3] & C[2]) or @C[2],@T[1] mov $A[2][3](%rdi),@C[2] xor @T[0],@T[1] # C[0] ^ (C[1] | C[2]) mov $A[3][4](%rdi),@C[3] mov @T[1],$A[1][0](%rsi) # R[1][0] = C[0] ^ (C[1] | C[2]) xor @D[3],@C[2] xor @D[4],@C[3] rol \$$rhotates[2][3],@C[2] xor @D[2],@C[1] rol \$$rhotates[3][4],@C[3] xor @D[0],@C[4] rol \$$rhotates[1][2],@C[1] xor @D[1],@C[0] rol \$$rhotates[4][0],@C[4] mov @C[2],@T[0] and @C[3],@C[2] rol \$$rhotates[0][1],@C[0] not @C[3] xor @C[1],@C[2] # C[1] ^ ( C[2] & C[3]) mov @C[2],$A[2][1](%rsi) # R[2][1] = C[1] ^ ( C[2] & C[3]) mov @C[4],@T[1] and @C[3],@C[4] mov $A[2][1](%rdi),@C[2] xor @T[0],@C[4] # C[2] ^ ( C[4] & ~C[3]) mov @C[4],$A[2][2](%rsi) # R[2][2] = C[2] ^ ( C[4] & ~C[3]) or @C[1],@T[0] mov $A[4][3](%rdi),@C[4] xor @C[0],@T[0] # C[0] ^ ( C[2] | C[1]) mov @T[0],$A[2][0](%rsi) # R[2][0] = C[0] ^ ( C[2] | C[1]) and @C[0],@C[1] xor @T[1],@C[1] # C[4] ^ ( C[1] & C[0]) mov @C[1],$A[2][4](%rsi) # R[2][4] = C[4] ^ ( C[1] & C[0]) or @C[0],@T[1] mov $A[1][0](%rdi),@C[1] xor @C[3],@T[1] # ~C[3] ^ ( C[0] | C[4]) mov $A[3][2](%rdi),@C[3] mov @T[1],$A[2][3](%rsi) # R[2][3] = ~C[3] ^ ( C[0] | C[4]) mov $A[0][4](%rdi),@C[0] xor @D[1],@C[2] xor @D[2],@C[3] rol \$$rhotates[2][1],@C[2] xor @D[0],@C[1] rol \$$rhotates[3][2],@C[3] xor @D[3],@C[4] rol \$$rhotates[1][0],@C[1] xor @D[4],@C[0] rol \$$rhotates[4][3],@C[4] mov @C[2],@T[0] or @C[3],@C[2] rol \$$rhotates[0][4],@C[0] not @C[3] xor @C[1],@C[2] # C[1] ^ ( C[2] | C[3]) mov @C[2],$A[3][1](%rsi) # R[3][1] = C[1] ^ ( C[2] | C[3]) mov @C[4],@T[1] or @C[3],@C[4] xor @T[0],@C[4] # C[2] ^ ( C[4] | ~C[3]) mov @C[4],$A[3][2](%rsi) # R[3][2] = C[2] ^ ( C[4] | ~C[3]) and @C[1],@T[0] xor @C[0],@T[0] # C[0] ^ ( C[2] & C[1]) mov @T[0],$A[3][0](%rsi) # R[3][0] = C[0] ^ ( C[2] & C[1]) or @C[0],@C[1] xor @T[1],@C[1] # C[4] ^ ( C[1] | C[0]) mov @C[1],$A[3][4](%rsi) # R[3][4] = C[4] ^ ( C[1] | C[0]) and @T[1],@C[0] xor @C[3],@C[0] # ~C[3] ^ ( C[0] & C[4]) mov @C[0],$A[3][3](%rsi) # R[3][3] = ~C[3] ^ ( C[0] & C[4]) xor $A[0][2](%rdi),@D[2] xor $A[1][3](%rdi),@D[3] rol \$$rhotates[0][2],@D[2] xor $A[4][1](%rdi),@D[1] rol \$$rhotates[1][3],@D[3] xor $A[2][4](%rdi),@D[4] rol \$$rhotates[4][1],@D[1] xor $A[3][0](%rdi),@D[0] xchg %rsi,%rdi rol \$$rhotates[2][4],@D[4] rol \$$rhotates[3][0],@D[0] ___ @C = @D[2..4,0,1]; $code.=<<___; mov @C[0],@T[0] and @C[1],@C[0] not @C[1] xor @C[4],@C[0] # C[4] ^ ( C[0] & C[1]) mov @C[0],$A[4][4](%rdi) # R[4][4] = C[4] ^ ( C[0] & C[1]) mov @C[2],@T[1] and @C[1],@C[2] xor @T[0],@C[2] # C[0] ^ ( C[2] & ~C[1]) mov @C[2],$A[4][0](%rdi) # R[4][0] = C[0] ^ ( C[2] & ~C[1]) or @C[4],@T[0] xor @C[3],@T[0] # C[3] ^ ( C[0] | C[4]) mov @T[0],$A[4][3](%rdi) # R[4][3] = C[3] ^ ( C[0] | C[4]) and @C[3],@C[4] xor @T[1],@C[4] # C[2] ^ ( C[4] & C[3]) mov @C[4],$A[4][2](%rdi) # R[4][2] = C[2] ^ ( C[4] & C[3]) or @T[1],@C[3] xor @C[1],@C[3] # ~C[1] ^ ( C[2] | C[3]) mov @C[3],$A[4][1](%rdi) # R[4][1] = ~C[1] ^ ( C[2] | C[3]) mov @C[0],@C[1] # harmonize with the loop top mov @T[0],@C[0] test \$255,$iotas jnz .Loop lea -192($iotas),$iotas # rewind iotas ret .size __KeccakF1600,.-__KeccakF1600 .globl KeccakF1600 .type KeccakF1600,\@function,1,"unwind" .align 32 KeccakF1600: .cfi_startproc push %rbx .cfi_push %rbx push %rbp .cfi_push %rbp push %r12 .cfi_push %r12 push %r13 .cfi_push %r13 push %r14 .cfi_push %r14 push %r15 .cfi_push %r15 lea 100(%rdi),%rdi # size optimization sub \$200,%rsp .cfi_alloca 200 .cfi_end_prologue notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea iotas(%rip),$iotas lea 100(%rsp),%rsi # size optimization call __KeccakF1600 notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea -100(%rdi),%rdi # preserve A[][] lea 248(%rsp),%r11 .cfi_def_cfa %r11,8 mov -48(%r11),%r15 mov -40(%r11),%r14 mov -32(%r11),%r13 mov -24(%r11),%r12 mov -16(%r11),%rbp mov -8(%r11),%rbx lea (%r11),%rsp .cfi_epilogue ret .cfi_endproc .size KeccakF1600,.-KeccakF1600 ___ { my ($A_flat,$inp,$len,$bsz) = ("%rdi","%rsi","%rdx","%rcx"); ($A_flat,$inp) = ("%r8","%r9"); $code.=<<___; .globl SHA3_absorb .type SHA3_absorb,\@function,4,"unwind" .align 32 SHA3_absorb: .cfi_startproc push %rbx .cfi_push %rbx push %rbp .cfi_push %rbp push %r12 .cfi_push %r12 push %r13 .cfi_push %r13 push %r14 .cfi_push %r14 push %r15 .cfi_push %r15 lea 100(%rdi),%rdi # size optimization sub \$232,%rsp .cfi_alloca 232 .cfi_end_prologue mov %rsi,$inp lea 100(%rsp),%rsi # size optimization notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea iotas(%rip),$iotas mov $bsz,216-100(%rsi) # save bsz .Loop_absorb: cmp $bsz,$len jc .Ldone_absorb shr \$3,$bsz lea -100(%rdi),$A_flat .Lblock_absorb: mov ($inp),%rax lea 8($inp),$inp xor ($A_flat),%rax lea 8($A_flat),$A_flat sub \$8,$len mov %rax,-8($A_flat) sub \$1,$bsz jnz .Lblock_absorb mov $inp,200-100(%rsi) # save inp mov $len,208-100(%rsi) # save len call __KeccakF1600 mov 200-100(%rsi),$inp # pull inp mov 208-100(%rsi),$len # pull len mov 216-100(%rsi),$bsz # pull bsz jmp .Loop_absorb .align 32 .Ldone_absorb: mov $len,%rax # return value notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea 280(%rsp),%r11 .cfi_def_cfa %r11,8 mov -48(%r11),%r15 mov -40(%r11),%r14 mov -32(%r11),%r13 mov -24(%r11),%r12 mov -16(%r11),%rbp mov -8(%r11),%rbx lea (%r11),%rsp .cfi_epilogue ret .cfi_endproc .size SHA3_absorb,.-SHA3_absorb ___ } { my ($A_flat,$out,$len,$bsz) = ("%rdi","%rsi","%rdx","%rcx"); ($out,$len,$bsz) = ("%r12","%r13","%r14"); $code.=<<___; .globl SHA3_squeeze .type SHA3_squeeze,\@function,4,"unwind" .align 32 SHA3_squeeze: .cfi_startproc push %r12 .cfi_push %r12 push %r13 .cfi_push %r13 push %r14 .cfi_push %r14 sub \$32,%rsp # Windows thing .cfi_alloca 32 .cfi_end_prologue shr \$3,%rcx mov $A_flat,%r8 mov %rsi,$out mov %rdx,$len mov %rcx,$bsz jmp .Loop_squeeze .align 32 .Loop_squeeze: cmp \$8,$len jb .Ltail_squeeze mov (%r8),%rax lea 8(%r8),%r8 mov %rax,($out) lea 8($out),$out sub \$8,$len # len -= 8 jz .Ldone_squeeze sub \$1,%rcx # bsz-- jnz .Loop_squeeze mov %rdi,%rcx # Windows thing call KeccakF1600 mov $A_flat,%r8 mov $bsz,%rcx jmp .Loop_squeeze .Ltail_squeeze: mov %r8, %rsi mov $out,%rdi mov $len,%rcx .byte 0xf3,0xa4 # rep movsb .Ldone_squeeze: mov 32(%rsp),%r14 mov 40(%rsp),%r13 mov 48(%rsp),%r12 add \$56,%rsp .cfi_alloca -56 .cfi_epilogue ret .cfi_endproc .size SHA3_squeeze,.-SHA3_squeeze ___ } $code.=<<___; .align 256 .quad 0,0,0,0,0,0,0,0 .type iotas,\@object iotas: .quad 0x0000000000000001 .quad 0x0000000000008082 .quad 0x800000000000808a .quad 0x8000000080008000 .quad 0x000000000000808b .quad 0x0000000080000001 .quad 0x8000000080008081 .quad 0x8000000000008009 .quad 0x000000000000008a .quad 0x0000000000000088 .quad 0x0000000080008009 .quad 0x000000008000000a .quad 0x000000008000808b .quad 0x800000000000008b .quad 0x8000000000008089 .quad 0x8000000000008003 .quad 0x8000000000008002 .quad 0x8000000000000080 .quad 0x000000000000800a .quad 0x800000008000000a .quad 0x8000000080008081 .quad 0x8000000000008080 .quad 0x0000000080000001 .quad 0x8000000080008008 .size iotas,.-iotas .asciz "Keccak-1600 absorb and squeeze for x86_64, CRYPTOGAMS by " ___ foreach (split("\n",$code)) { # Below replacement results in 11.2 on Sandy Bridge, 9.4 on # Haswell, but it hurts other processors by up to 2-3-4x... #s/rol\s+(\$[0-9]+),(%[a-z][a-z0-9]+)/shld\t$1,$2,$2/; # Below replacement results in 9.3 on Haswell [as well as # on Ryzen, i.e. it *hurts* Ryzen]... #s/rol\s+\$([0-9]+),(%[a-z][a-z0-9]+)/rorx\t\$64-$1,$2,$2/; print $_, "\n"; } close STDOUT;