| /* |
| * x86_64/AVX2 assembler optimized version of Twofish |
| * |
| * Copyright © 2012-2013 Jussi Kivilinna <jussi.kivilinna@iki.fi> |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation; either version 2 of the License, or |
| * (at your option) any later version. |
| * |
| */ |
| |
| #include <linux/linkage.h> |
| #include "glue_helper-asm-avx2.S" |
| |
| .file "twofish-avx2-asm_64.S" |
| |
| .data |
| .align 16 |
| |
| .Lvpshufb_mask0: |
| .long 0x80808000 |
| .long 0x80808004 |
| .long 0x80808008 |
| .long 0x8080800c |
| |
| .Lbswap128_mask: |
| .byte 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0 |
| .Lxts_gf128mul_and_shl1_mask_0: |
| .byte 0x87, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0 |
| .Lxts_gf128mul_and_shl1_mask_1: |
| .byte 0x0e, 1, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0, 0, 0 |
| |
| .text |
| |
| /* structure of crypto context */ |
| #define s0 0 |
| #define s1 1024 |
| #define s2 2048 |
| #define s3 3072 |
| #define w 4096 |
| #define k 4128 |
| |
| /* register macros */ |
| #define CTX %rdi |
| |
| #define RS0 CTX |
| #define RS1 %r8 |
| #define RS2 %r9 |
| #define RS3 %r10 |
| #define RK %r11 |
| #define RW %rax |
| #define RROUND %r12 |
| #define RROUNDd %r12d |
| |
| #define RA0 %ymm8 |
| #define RB0 %ymm9 |
| #define RC0 %ymm10 |
| #define RD0 %ymm11 |
| #define RA1 %ymm12 |
| #define RB1 %ymm13 |
| #define RC1 %ymm14 |
| #define RD1 %ymm15 |
| |
| /* temp regs */ |
| #define RX0 %ymm0 |
| #define RY0 %ymm1 |
| #define RX1 %ymm2 |
| #define RY1 %ymm3 |
| #define RT0 %ymm4 |
| #define RIDX %ymm5 |
| |
| #define RX0x %xmm0 |
| #define RY0x %xmm1 |
| #define RX1x %xmm2 |
| #define RY1x %xmm3 |
| #define RT0x %xmm4 |
| |
| /* vpgatherdd mask and '-1' */ |
| #define RNOT %ymm6 |
| |
| /* byte mask, (-1 >> 24) */ |
| #define RBYTE %ymm7 |
| |
| /********************************************************************** |
| 16-way AVX2 twofish |
| **********************************************************************/ |
| #define init_round_constants() \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpsrld $24, RNOT, RBYTE; \ |
| leaq k(CTX), RK; \ |
| leaq w(CTX), RW; \ |
| leaq s1(CTX), RS1; \ |
| leaq s2(CTX), RS2; \ |
| leaq s3(CTX), RS3; \ |
| |
| #define g16(ab, rs0, rs1, rs2, rs3, xy) \ |
| vpand RBYTE, ab ## 0, RIDX; \ |
| vpgatherdd RNOT, (rs0, RIDX, 4), xy ## 0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| \ |
| vpand RBYTE, ab ## 1, RIDX; \ |
| vpgatherdd RNOT, (rs0, RIDX, 4), xy ## 1; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| \ |
| vpsrld $8, ab ## 0, RIDX; \ |
| vpand RBYTE, RIDX, RIDX; \ |
| vpgatherdd RNOT, (rs1, RIDX, 4), RT0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpxor RT0, xy ## 0, xy ## 0; \ |
| \ |
| vpsrld $8, ab ## 1, RIDX; \ |
| vpand RBYTE, RIDX, RIDX; \ |
| vpgatherdd RNOT, (rs1, RIDX, 4), RT0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpxor RT0, xy ## 1, xy ## 1; \ |
| \ |
| vpsrld $16, ab ## 0, RIDX; \ |
| vpand RBYTE, RIDX, RIDX; \ |
| vpgatherdd RNOT, (rs2, RIDX, 4), RT0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpxor RT0, xy ## 0, xy ## 0; \ |
| \ |
| vpsrld $16, ab ## 1, RIDX; \ |
| vpand RBYTE, RIDX, RIDX; \ |
| vpgatherdd RNOT, (rs2, RIDX, 4), RT0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpxor RT0, xy ## 1, xy ## 1; \ |
| \ |
| vpsrld $24, ab ## 0, RIDX; \ |
| vpgatherdd RNOT, (rs3, RIDX, 4), RT0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpxor RT0, xy ## 0, xy ## 0; \ |
| \ |
| vpsrld $24, ab ## 1, RIDX; \ |
| vpgatherdd RNOT, (rs3, RIDX, 4), RT0; \ |
| vpcmpeqd RNOT, RNOT, RNOT; \ |
| vpxor RT0, xy ## 1, xy ## 1; |
| |
| #define g1_16(a, x) \ |
| g16(a, RS0, RS1, RS2, RS3, x); |
| |
| #define g2_16(b, y) \ |
| g16(b, RS1, RS2, RS3, RS0, y); |
| |
| #define encrypt_round_end16(a, b, c, d, nk) \ |
| vpaddd RY0, RX0, RX0; \ |
| vpaddd RX0, RY0, RY0; \ |
| vpbroadcastd nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RX0, RX0; \ |
| vpbroadcastd 4+nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RY0, RY0; \ |
| \ |
| vpxor RY0, d ## 0, d ## 0; \ |
| \ |
| vpxor RX0, c ## 0, c ## 0; \ |
| vpsrld $1, c ## 0, RT0; \ |
| vpslld $31, c ## 0, c ## 0; \ |
| vpor RT0, c ## 0, c ## 0; \ |
| \ |
| vpaddd RY1, RX1, RX1; \ |
| vpaddd RX1, RY1, RY1; \ |
| vpbroadcastd nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RX1, RX1; \ |
| vpbroadcastd 4+nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RY1, RY1; \ |
| \ |
| vpxor RY1, d ## 1, d ## 1; \ |
| \ |
| vpxor RX1, c ## 1, c ## 1; \ |
| vpsrld $1, c ## 1, RT0; \ |
| vpslld $31, c ## 1, c ## 1; \ |
| vpor RT0, c ## 1, c ## 1; \ |
| |
| #define encrypt_round16(a, b, c, d, nk) \ |
| g2_16(b, RY); \ |
| \ |
| vpslld $1, b ## 0, RT0; \ |
| vpsrld $31, b ## 0, b ## 0; \ |
| vpor RT0, b ## 0, b ## 0; \ |
| \ |
| vpslld $1, b ## 1, RT0; \ |
| vpsrld $31, b ## 1, b ## 1; \ |
| vpor RT0, b ## 1, b ## 1; \ |
| \ |
| g1_16(a, RX); \ |
| \ |
| encrypt_round_end16(a, b, c, d, nk); |
| |
| #define encrypt_round_first16(a, b, c, d, nk) \ |
| vpslld $1, d ## 0, RT0; \ |
| vpsrld $31, d ## 0, d ## 0; \ |
| vpor RT0, d ## 0, d ## 0; \ |
| \ |
| vpslld $1, d ## 1, RT0; \ |
| vpsrld $31, d ## 1, d ## 1; \ |
| vpor RT0, d ## 1, d ## 1; \ |
| \ |
| encrypt_round16(a, b, c, d, nk); |
| |
| #define encrypt_round_last16(a, b, c, d, nk) \ |
| g2_16(b, RY); \ |
| \ |
| g1_16(a, RX); \ |
| \ |
| encrypt_round_end16(a, b, c, d, nk); |
| |
| #define decrypt_round_end16(a, b, c, d, nk) \ |
| vpaddd RY0, RX0, RX0; \ |
| vpaddd RX0, RY0, RY0; \ |
| vpbroadcastd nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RX0, RX0; \ |
| vpbroadcastd 4+nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RY0, RY0; \ |
| \ |
| vpxor RX0, c ## 0, c ## 0; \ |
| \ |
| vpxor RY0, d ## 0, d ## 0; \ |
| vpsrld $1, d ## 0, RT0; \ |
| vpslld $31, d ## 0, d ## 0; \ |
| vpor RT0, d ## 0, d ## 0; \ |
| \ |
| vpaddd RY1, RX1, RX1; \ |
| vpaddd RX1, RY1, RY1; \ |
| vpbroadcastd nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RX1, RX1; \ |
| vpbroadcastd 4+nk(RK,RROUND,8), RT0; \ |
| vpaddd RT0, RY1, RY1; \ |
| \ |
| vpxor RX1, c ## 1, c ## 1; \ |
| \ |
| vpxor RY1, d ## 1, d ## 1; \ |
| vpsrld $1, d ## 1, RT0; \ |
| vpslld $31, d ## 1, d ## 1; \ |
| vpor RT0, d ## 1, d ## 1; |
| |
| #define decrypt_round16(a, b, c, d, nk) \ |
| g1_16(a, RX); \ |
| \ |
| vpslld $1, a ## 0, RT0; \ |
| vpsrld $31, a ## 0, a ## 0; \ |
| vpor RT0, a ## 0, a ## 0; \ |
| \ |
| vpslld $1, a ## 1, RT0; \ |
| vpsrld $31, a ## 1, a ## 1; \ |
| vpor RT0, a ## 1, a ## 1; \ |
| \ |
| g2_16(b, RY); \ |
| \ |
| decrypt_round_end16(a, b, c, d, nk); |
| |
| #define decrypt_round_first16(a, b, c, d, nk) \ |
| vpslld $1, c ## 0, RT0; \ |
| vpsrld $31, c ## 0, c ## 0; \ |
| vpor RT0, c ## 0, c ## 0; \ |
| \ |
| vpslld $1, c ## 1, RT0; \ |
| vpsrld $31, c ## 1, c ## 1; \ |
| vpor RT0, c ## 1, c ## 1; \ |
| \ |
| decrypt_round16(a, b, c, d, nk) |
| |
| #define decrypt_round_last16(a, b, c, d, nk) \ |
| g1_16(a, RX); \ |
| \ |
| g2_16(b, RY); \ |
| \ |
| decrypt_round_end16(a, b, c, d, nk); |
| |
| #define encrypt_cycle16() \ |
| encrypt_round16(RA, RB, RC, RD, 0); \ |
| encrypt_round16(RC, RD, RA, RB, 8); |
| |
| #define encrypt_cycle_first16() \ |
| encrypt_round_first16(RA, RB, RC, RD, 0); \ |
| encrypt_round16(RC, RD, RA, RB, 8); |
| |
| #define encrypt_cycle_last16() \ |
| encrypt_round16(RA, RB, RC, RD, 0); \ |
| encrypt_round_last16(RC, RD, RA, RB, 8); |
| |
| #define decrypt_cycle16(n) \ |
| decrypt_round16(RC, RD, RA, RB, 8); \ |
| decrypt_round16(RA, RB, RC, RD, 0); |
| |
| #define decrypt_cycle_first16(n) \ |
| decrypt_round_first16(RC, RD, RA, RB, 8); \ |
| decrypt_round16(RA, RB, RC, RD, 0); |
| |
| #define decrypt_cycle_last16(n) \ |
| decrypt_round16(RC, RD, RA, RB, 8); \ |
| decrypt_round_last16(RA, RB, RC, RD, 0); |
| |
| #define transpose_4x4(x0,x1,x2,x3,t1,t2) \ |
| vpunpckhdq x1, x0, t2; \ |
| vpunpckldq x1, x0, x0; \ |
| \ |
| vpunpckldq x3, x2, t1; \ |
| vpunpckhdq x3, x2, x2; \ |
| \ |
| vpunpckhqdq t1, x0, x1; \ |
| vpunpcklqdq t1, x0, x0; \ |
| \ |
| vpunpckhqdq x2, t2, x3; \ |
| vpunpcklqdq x2, t2, x2; |
| |
| #define read_blocks8(offs,a,b,c,d) \ |
| transpose_4x4(a, b, c, d, RX0, RY0); |
| |
| #define write_blocks8(offs,a,b,c,d) \ |
| transpose_4x4(a, b, c, d, RX0, RY0); |
| |
| #define inpack_enc8(a,b,c,d) \ |
| vpbroadcastd 4*0(RW), RT0; \ |
| vpxor RT0, a, a; \ |
| \ |
| vpbroadcastd 4*1(RW), RT0; \ |
| vpxor RT0, b, b; \ |
| \ |
| vpbroadcastd 4*2(RW), RT0; \ |
| vpxor RT0, c, c; \ |
| \ |
| vpbroadcastd 4*3(RW), RT0; \ |
| vpxor RT0, d, d; |
| |
| #define outunpack_enc8(a,b,c,d) \ |
| vpbroadcastd 4*4(RW), RX0; \ |
| vpbroadcastd 4*5(RW), RY0; \ |
| vpxor RX0, c, RX0; \ |
| vpxor RY0, d, RY0; \ |
| \ |
| vpbroadcastd 4*6(RW), RT0; \ |
| vpxor RT0, a, c; \ |
| vpbroadcastd 4*7(RW), RT0; \ |
| vpxor RT0, b, d; \ |
| \ |
| vmovdqa RX0, a; \ |
| vmovdqa RY0, b; |
| |
| #define inpack_dec8(a,b,c,d) \ |
| vpbroadcastd 4*4(RW), RX0; \ |
| vpbroadcastd 4*5(RW), RY0; \ |
| vpxor RX0, a, RX0; \ |
| vpxor RY0, b, RY0; \ |
| \ |
| vpbroadcastd 4*6(RW), RT0; \ |
| vpxor RT0, c, a; \ |
| vpbroadcastd 4*7(RW), RT0; \ |
| vpxor RT0, d, b; \ |
| \ |
| vmovdqa RX0, c; \ |
| vmovdqa RY0, d; |
| |
| #define outunpack_dec8(a,b,c,d) \ |
| vpbroadcastd 4*0(RW), RT0; \ |
| vpxor RT0, a, a; \ |
| \ |
| vpbroadcastd 4*1(RW), RT0; \ |
| vpxor RT0, b, b; \ |
| \ |
| vpbroadcastd 4*2(RW), RT0; \ |
| vpxor RT0, c, c; \ |
| \ |
| vpbroadcastd 4*3(RW), RT0; \ |
| vpxor RT0, d, d; |
| |
| #define read_blocks16(a,b,c,d) \ |
| read_blocks8(0, a ## 0, b ## 0, c ## 0, d ## 0); \ |
| read_blocks8(8, a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| #define write_blocks16(a,b,c,d) \ |
| write_blocks8(0, a ## 0, b ## 0, c ## 0, d ## 0); \ |
| write_blocks8(8, a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| #define xor_blocks16(a,b,c,d) \ |
| xor_blocks8(0, a ## 0, b ## 0, c ## 0, d ## 0); \ |
| xor_blocks8(8, a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| #define inpack_enc16(a,b,c,d) \ |
| inpack_enc8(a ## 0, b ## 0, c ## 0, d ## 0); \ |
| inpack_enc8(a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| #define outunpack_enc16(a,b,c,d) \ |
| outunpack_enc8(a ## 0, b ## 0, c ## 0, d ## 0); \ |
| outunpack_enc8(a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| #define inpack_dec16(a,b,c,d) \ |
| inpack_dec8(a ## 0, b ## 0, c ## 0, d ## 0); \ |
| inpack_dec8(a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| #define outunpack_dec16(a,b,c,d) \ |
| outunpack_dec8(a ## 0, b ## 0, c ## 0, d ## 0); \ |
| outunpack_dec8(a ## 1, b ## 1, c ## 1, d ## 1); |
| |
| .align 8 |
| __twofish_enc_blk16: |
| /* input: |
| * %rdi: ctx, CTX |
| * RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: plaintext |
| * output: |
| * RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: ciphertext |
| */ |
| init_round_constants(); |
| |
| read_blocks16(RA, RB, RC, RD); |
| inpack_enc16(RA, RB, RC, RD); |
| |
| xorl RROUNDd, RROUNDd; |
| encrypt_cycle_first16(); |
| movl $2, RROUNDd; |
| |
| .align 4 |
| .L__enc_loop: |
| encrypt_cycle16(); |
| |
| addl $2, RROUNDd; |
| cmpl $14, RROUNDd; |
| jne .L__enc_loop; |
| |
| encrypt_cycle_last16(); |
| |
| outunpack_enc16(RA, RB, RC, RD); |
| write_blocks16(RA, RB, RC, RD); |
| |
| ret; |
| ENDPROC(__twofish_enc_blk16) |
| |
| .align 8 |
| __twofish_dec_blk16: |
| /* input: |
| * %rdi: ctx, CTX |
| * RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: ciphertext |
| * output: |
| * RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: plaintext |
| */ |
| init_round_constants(); |
| |
| read_blocks16(RA, RB, RC, RD); |
| inpack_dec16(RA, RB, RC, RD); |
| |
| movl $14, RROUNDd; |
| decrypt_cycle_first16(); |
| movl $12, RROUNDd; |
| |
| .align 4 |
| .L__dec_loop: |
| decrypt_cycle16(); |
| |
| addl $-2, RROUNDd; |
| jnz .L__dec_loop; |
| |
| decrypt_cycle_last16(); |
| |
| outunpack_dec16(RA, RB, RC, RD); |
| write_blocks16(RA, RB, RC, RD); |
| |
| ret; |
| ENDPROC(__twofish_dec_blk16) |
| |
| ENTRY(twofish_ecb_enc_16way) |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst |
| * %rdx: src |
| */ |
| |
| vzeroupper; |
| pushq %r12; |
| |
| load_16way(%rdx, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| call __twofish_enc_blk16; |
| |
| store_16way(%rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| popq %r12; |
| vzeroupper; |
| |
| ret; |
| ENDPROC(twofish_ecb_enc_16way) |
| |
| ENTRY(twofish_ecb_dec_16way) |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst |
| * %rdx: src |
| */ |
| |
| vzeroupper; |
| pushq %r12; |
| |
| load_16way(%rdx, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| call __twofish_dec_blk16; |
| |
| store_16way(%rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| popq %r12; |
| vzeroupper; |
| |
| ret; |
| ENDPROC(twofish_ecb_dec_16way) |
| |
| ENTRY(twofish_cbc_dec_16way) |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst |
| * %rdx: src |
| */ |
| |
| vzeroupper; |
| pushq %r12; |
| |
| load_16way(%rdx, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| call __twofish_dec_blk16; |
| |
| store_cbc_16way(%rdx, %rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1, |
| RX0); |
| |
| popq %r12; |
| vzeroupper; |
| |
| ret; |
| ENDPROC(twofish_cbc_dec_16way) |
| |
| ENTRY(twofish_ctr_16way) |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst (16 blocks) |
| * %rdx: src (16 blocks) |
| * %rcx: iv (little endian, 128bit) |
| */ |
| |
| vzeroupper; |
| pushq %r12; |
| |
| load_ctr_16way(%rcx, .Lbswap128_mask, RA0, RB0, RC0, RD0, RA1, RB1, RC1, |
| RD1, RX0, RX0x, RX1, RX1x, RY0, RY0x, RY1, RY1x, RNOT, |
| RBYTE); |
| |
| call __twofish_enc_blk16; |
| |
| store_ctr_16way(%rdx, %rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| popq %r12; |
| vzeroupper; |
| |
| ret; |
| ENDPROC(twofish_ctr_16way) |
| |
| .align 8 |
| twofish_xts_crypt_16way: |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst (16 blocks) |
| * %rdx: src (16 blocks) |
| * %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸)) |
| * %r8: pointer to __twofish_enc_blk16 or __twofish_dec_blk16 |
| */ |
| |
| vzeroupper; |
| pushq %r12; |
| |
| load_xts_16way(%rcx, %rdx, %rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, |
| RD1, RX0, RX0x, RX1, RX1x, RY0, RY0x, RY1, RY1x, RNOT, |
| .Lxts_gf128mul_and_shl1_mask_0, |
| .Lxts_gf128mul_and_shl1_mask_1); |
| |
| call *%r8; |
| |
| store_xts_16way(%rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1); |
| |
| popq %r12; |
| vzeroupper; |
| |
| ret; |
| ENDPROC(twofish_xts_crypt_16way) |
| |
| ENTRY(twofish_xts_enc_16way) |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst (16 blocks) |
| * %rdx: src (16 blocks) |
| * %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸)) |
| */ |
| leaq __twofish_enc_blk16, %r8; |
| jmp twofish_xts_crypt_16way; |
| ENDPROC(twofish_xts_enc_16way) |
| |
| ENTRY(twofish_xts_dec_16way) |
| /* input: |
| * %rdi: ctx, CTX |
| * %rsi: dst (16 blocks) |
| * %rdx: src (16 blocks) |
| * %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸)) |
| */ |
| leaq __twofish_dec_blk16, %r8; |
| jmp twofish_xts_crypt_16way; |
| ENDPROC(twofish_xts_dec_16way) |