📄 twofish-x86_64-asm_64.s
字号:
/**************************************************************************** Copyright (C) 2006 by Joachim Fritschi, <jfritschi@freenet.de> ** ** This program is free software; you can redistribute it and/or modify ** it under the terms of the GNU General Public License as published by ** the Free Software Foundation; either version 2 of the License, or ** (at your option) any later version. ** ** This program is distributed in the hope that it will be useful, ** but WITHOUT ANY WARRANTY; without even the implied warranty of ** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ** GNU General Public License for more details. ** ** You should have received a copy of the GNU General Public License ** along with this program; if not, write to the ** Free Software Foundation, Inc., ** 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. ****************************************************************************/.file "twofish-x86_64-asm.S".text#include <asm/asm-offsets.h>#define a_offset 0#define b_offset 4#define c_offset 8#define d_offset 12/* Structure of the crypto context struct*/#define s0 0 /* S0 Array 256 Words each */#define s1 1024 /* S1 Array */#define s2 2048 /* S2 Array */#define s3 3072 /* S3 Array */#define w 4096 /* 8 whitening keys (word) */#define k 4128 /* key 1-32 ( word ) *//* define a few register aliases to allow macro substitution */#define R0 %rax#define R0D %eax#define R0B %al#define R0H %ah#define R1 %rbx#define R1D %ebx#define R1B %bl#define R1H %bh#define R2 %rcx#define R2D %ecx#define R2B %cl#define R2H %ch#define R3 %rdx#define R3D %edx#define R3B %dl#define R3H %dh/* performs input whitening */#define input_whitening(src,context,offset)\ xor w+offset(context), src;/* performs input whitening */#define output_whitening(src,context,offset)\ xor w+16+offset(context), src;/* * a input register containing a (rotated 16) * b input register containing b * c input register containing c * d input register containing d (already rol $1) * operations on a and b are interleaved to increase performance */#define encrypt_round(a,b,c,d,round)\ movzx b ## B, %edi;\ mov s1(%r11,%rdi,4),%r8d;\ movzx a ## B, %edi;\ mov s2(%r11,%rdi,4),%r9d;\ movzx b ## H, %edi;\ ror $16, b ## D;\ xor s2(%r11,%rdi,4),%r8d;\ movzx a ## H, %edi;\ ror $16, a ## D;\ xor s3(%r11,%rdi,4),%r9d;\ movzx b ## B, %edi;\ xor s3(%r11,%rdi,4),%r8d;\ movzx a ## B, %edi;\ xor (%r11,%rdi,4), %r9d;\ movzx b ## H, %edi;\ ror $15, b ## D;\ xor (%r11,%rdi,4), %r8d;\ movzx a ## H, %edi;\ xor s1(%r11,%rdi,4),%r9d;\ add %r8d, %r9d;\ add %r9d, %r8d;\ add k+round(%r11), %r9d;\ xor %r9d, c ## D;\ rol $15, c ## D;\ add k+4+round(%r11),%r8d;\ xor %r8d, d ## D;/* * a input register containing a(rotated 16) * b input register containing b * c input register containing c * d input register containing d (already rol $1) * operations on a and b are interleaved to increase performance * during the round a and b are prepared for the output whitening */#define encrypt_last_round(a,b,c,d,round)\ mov b ## D, %r10d;\ shl $32, %r10;\ movzx b ## B, %edi;\ mov s1(%r11,%rdi,4),%r8d;\ movzx a ## B, %edi;\ mov s2(%r11,%rdi,4),%r9d;\ movzx b ## H, %edi;\ ror $16, b ## D;\ xor s2(%r11,%rdi,4),%r8d;\ movzx a ## H, %edi;\ ror $16, a ## D;\ xor s3(%r11,%rdi,4),%r9d;\ movzx b ## B, %edi;\ xor s3(%r11,%rdi,4),%r8d;\ movzx a ## B, %edi;\ xor (%r11,%rdi,4), %r9d;\ xor a, %r10;\ movzx b ## H, %edi;\ xor (%r11,%rdi,4), %r8d;\ movzx a ## H, %edi;\ xor s1(%r11,%rdi,4),%r9d;\ add %r8d, %r9d;\ add %r9d, %r8d;\ add k+round(%r11), %r9d;\ xor %r9d, c ## D;\ ror $1, c ## D;\ add k+4+round(%r11),%r8d;\ xor %r8d, d ## D/* * a input register containing a * b input register containing b (rotated 16) * c input register containing c (already rol $1) * d input register containing d * operations on a and b are interleaved to increase performance */#define decrypt_round(a,b,c,d,round)\ movzx a ## B, %edi;\ mov (%r11,%rdi,4), %r9d;\ movzx b ## B, %edi;\ mov s3(%r11,%rdi,4),%r8d;\ movzx a ## H, %edi;\ ror $16, a ## D;\ xor s1(%r11,%rdi,4),%r9d;\ movzx b ## H, %edi;\ ror $16, b ## D;\ xor (%r11,%rdi,4), %r8d;\ movzx a ## B, %edi;\ xor s2(%r11,%rdi,4),%r9d;\ movzx b ## B, %edi;\ xor s1(%r11,%rdi,4),%r8d;\ movzx a ## H, %edi;\ ror $15, a ## D;\ xor s3(%r11,%rdi,4),%r9d;\ movzx b ## H, %edi;\ xor s2(%r11,%rdi,4),%r8d;\ add %r8d, %r9d;\ add %r9d, %r8d;\ add k+round(%r11), %r9d;\ xor %r9d, c ## D;\ add k+4+round(%r11),%r8d;\ xor %r8d, d ## D;\ rol $15, d ## D;/* * a input register containing a * b input register containing b * c input register containing c (already rol $1) * d input register containing d * operations on a and b are interleaved to increase performance * during the round a and b are prepared for the output whitening */#define decrypt_last_round(a,b,c,d,round)\ movzx a ## B, %edi;\ mov (%r11,%rdi,4), %r9d;\ movzx b ## B, %edi;\ mov s3(%r11,%rdi,4),%r8d;\ movzx b ## H, %edi;\ ror $16, b ## D;\ xor (%r11,%rdi,4), %r8d;\ movzx a ## H, %edi;\ mov b ## D, %r10d;\ shl $32, %r10;\ xor a, %r10;\ ror $16, a ## D;\ xor s1(%r11,%rdi,4),%r9d;\ movzx b ## B, %edi;\ xor s1(%r11,%rdi,4),%r8d;\ movzx a ## B, %edi;\ xor s2(%r11,%rdi,4),%r9d;\ movzx b ## H, %edi;\ xor s2(%r11,%rdi,4),%r8d;\ movzx a ## H, %edi;\ xor s3(%r11,%rdi,4),%r9d;\ add %r8d, %r9d;\ add %r9d, %r8d;\ add k+round(%r11), %r9d;\ xor %r9d, c ## D;\ add k+4+round(%r11),%r8d;\ xor %r8d, d ## D;\ ror $1, d ## D;.align 8.global twofish_enc_blk.global twofish_dec_blktwofish_enc_blk: pushq R1 /* %rdi contains the crypto tfm adress */ /* %rsi contains the output adress */ /* %rdx contains the input adress */ add $crypto_tfm_ctx_offset, %rdi /* set ctx adress */ /* ctx adress is moved to free one non-rex register as target for the 8bit high operations */ mov %rdi, %r11 movq (R3), R1 movq 8(R3), R3 input_whitening(R1,%r11,a_offset) input_whitening(R3,%r11,c_offset) mov R1D, R0D rol $16, R0D shr $32, R1 mov R3D, R2D shr $32, R3 rol $1, R3D encrypt_round(R0,R1,R2,R3,0); encrypt_round(R2,R3,R0,R1,8); encrypt_round(R0,R1,R2,R3,2*8); encrypt_round(R2,R3,R0,R1,3*8); encrypt_round(R0,R1,R2,R3,4*8); encrypt_round(R2,R3,R0,R1,5*8); encrypt_round(R0,R1,R2,R3,6*8); encrypt_round(R2,R3,R0,R1,7*8); encrypt_round(R0,R1,R2,R3,8*8); encrypt_round(R2,R3,R0,R1,9*8); encrypt_round(R0,R1,R2,R3,10*8); encrypt_round(R2,R3,R0,R1,11*8); encrypt_round(R0,R1,R2,R3,12*8); encrypt_round(R2,R3,R0,R1,13*8); encrypt_round(R0,R1,R2,R3,14*8); encrypt_last_round(R2,R3,R0,R1,15*8); output_whitening(%r10,%r11,a_offset) movq %r10, (%rsi) shl $32, R1 xor R0, R1 output_whitening(R1,%r11,c_offset) movq R1, 8(%rsi) popq R1 movq $1,%rax rettwofish_dec_blk: pushq R1 /* %rdi contains the crypto tfm adress */ /* %rsi contains the output adress */ /* %rdx contains the input adress */ add $crypto_tfm_ctx_offset, %rdi /* set ctx adress */ /* ctx adress is moved to free one non-rex register as target for the 8bit high operations */ mov %rdi, %r11 movq (R3), R1 movq 8(R3), R3 output_whitening(R1,%r11,a_offset) output_whitening(R3,%r11,c_offset) mov R1D, R0D shr $32, R1 rol $16, R1D mov R3D, R2D shr $32, R3 rol $1, R2D decrypt_round(R0,R1,R2,R3,15*8); decrypt_round(R2,R3,R0,R1,14*8); decrypt_round(R0,R1,R2,R3,13*8); decrypt_round(R2,R3,R0,R1,12*8); decrypt_round(R0,R1,R2,R3,11*8); decrypt_round(R2,R3,R0,R1,10*8); decrypt_round(R0,R1,R2,R3,9*8); decrypt_round(R2,R3,R0,R1,8*8); decrypt_round(R0,R1,R2,R3,7*8); decrypt_round(R2,R3,R0,R1,6*8); decrypt_round(R0,R1,R2,R3,5*8); decrypt_round(R2,R3,R0,R1,4*8); decrypt_round(R0,R1,R2,R3,3*8); decrypt_round(R2,R3,R0,R1,2*8); decrypt_round(R0,R1,R2,R3,1*8); decrypt_last_round(R2,R3,R0,R1,0); input_whitening(%r10,%r11,a_offset) movq %r10, (%rsi) shl $32, R1 xor R0, R1 input_whitening(R1,%r11,c_offset) movq R1, 8(%rsi) popq R1 movq $1,%rax ret
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -