mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-29 09:12:07 +00:00
8c8e40470f
Add an implementation of SHA-256 and SHA-224 using the Zvknha or Zvknhb extension. The assembly code is derived from OpenSSL code (openssl/openssl#21923) that was dual-licensed so that it could be reused in the kernel. Nevertheless, the assembly has been significantly reworked for integration with the kernel, for example by using a regular .S file instead of the so-called perlasm, using the assembler instead of bare '.inst', and greatly reducing code duplication. Co-developed-by: Charalampos Mitrodimas <charalampos.mitrodimas@vrull.eu> Signed-off-by: Charalampos Mitrodimas <charalampos.mitrodimas@vrull.eu> Co-developed-by: Heiko Stuebner <heiko.stuebner@vrull.eu> Signed-off-by: Heiko Stuebner <heiko.stuebner@vrull.eu> Co-developed-by: Phoebe Chen <phoebe.chen@sifive.com> Signed-off-by: Phoebe Chen <phoebe.chen@sifive.com> Signed-off-by: Jerry Shih <jerry.shih@sifive.com> Co-developed-by: Eric Biggers <ebiggers@google.com> Signed-off-by: Eric Biggers <ebiggers@google.com> Link: https://lore.kernel.org/r/20240122002024.27477-8-ebiggers@kernel.org Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
226 lines
7.0 KiB
ArmAsm
226 lines
7.0 KiB
ArmAsm
/* SPDX-License-Identifier: Apache-2.0 OR BSD-2-Clause */
|
|
//
|
|
// This file is dual-licensed, meaning that you can use it under your
|
|
// choice of either of the following two licenses:
|
|
//
|
|
// Copyright 2023 The OpenSSL Project Authors. All Rights Reserved.
|
|
//
|
|
// Licensed under the Apache License 2.0 (the "License"). You can obtain
|
|
// a copy in the file LICENSE in the source distribution or at
|
|
// https://www.openssl.org/source/license.html
|
|
//
|
|
// or
|
|
//
|
|
// Copyright (c) 2023, Christoph Müllner <christoph.muellner@vrull.eu>
|
|
// Copyright (c) 2023, Phoebe Chen <phoebe.chen@sifive.com>
|
|
// Copyright 2024 Google LLC
|
|
// All rights reserved.
|
|
//
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions
|
|
// are met:
|
|
// 1. Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// 2. Redistributions in binary form must reproduce the above copyright
|
|
// notice, this list of conditions and the following disclaimer in the
|
|
// documentation and/or other materials provided with the distribution.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
// The generated code of this file depends on the following RISC-V extensions:
|
|
// - RV64I
|
|
// - RISC-V Vector ('V') with VLEN >= 128
|
|
// - RISC-V Vector SHA-2 Secure Hash extension ('Zvknha' or 'Zvknhb')
|
|
// - RISC-V Vector Cryptography Bit-manipulation extension ('Zvkb')
|
|
|
|
#include <linux/cfi_types.h>
|
|
|
|
.text
|
|
.option arch, +zvknha, +zvkb
|
|
|
|
#define STATEP a0
|
|
#define DATA a1
|
|
#define NUM_BLOCKS a2
|
|
|
|
#define STATEP_C a3
|
|
|
|
#define MASK v0
|
|
#define INDICES v1
|
|
#define W0 v2
|
|
#define W1 v3
|
|
#define W2 v4
|
|
#define W3 v5
|
|
#define VTMP v6
|
|
#define FEBA v7
|
|
#define HGDC v8
|
|
#define K0 v10
|
|
#define K1 v11
|
|
#define K2 v12
|
|
#define K3 v13
|
|
#define K4 v14
|
|
#define K5 v15
|
|
#define K6 v16
|
|
#define K7 v17
|
|
#define K8 v18
|
|
#define K9 v19
|
|
#define K10 v20
|
|
#define K11 v21
|
|
#define K12 v22
|
|
#define K13 v23
|
|
#define K14 v24
|
|
#define K15 v25
|
|
#define PREV_FEBA v26
|
|
#define PREV_HGDC v27
|
|
|
|
// Do 4 rounds of SHA-256. w0 contains the current 4 message schedule words.
|
|
//
|
|
// If not all the message schedule words have been computed yet, then this also
|
|
// computes 4 more message schedule words. w1-w3 contain the next 3 groups of 4
|
|
// message schedule words; this macro computes the group after w3 and writes it
|
|
// to w0. This means that the next (w0, w1, w2, w3) is the current (w1, w2, w3,
|
|
// w0), so the caller must cycle through the registers accordingly.
|
|
.macro sha256_4rounds last, k, w0, w1, w2, w3
|
|
vadd.vv VTMP, \k, \w0
|
|
vsha2cl.vv HGDC, FEBA, VTMP
|
|
vsha2ch.vv FEBA, HGDC, VTMP
|
|
.if !\last
|
|
vmerge.vvm VTMP, \w2, \w1, MASK
|
|
vsha2ms.vv \w0, VTMP, \w3
|
|
.endif
|
|
.endm
|
|
|
|
.macro sha256_16rounds last, k0, k1, k2, k3
|
|
sha256_4rounds \last, \k0, W0, W1, W2, W3
|
|
sha256_4rounds \last, \k1, W1, W2, W3, W0
|
|
sha256_4rounds \last, \k2, W2, W3, W0, W1
|
|
sha256_4rounds \last, \k3, W3, W0, W1, W2
|
|
.endm
|
|
|
|
// void sha256_transform_zvknha_or_zvknhb_zvkb(u32 state[8], const u8 *data,
|
|
// int num_blocks);
|
|
SYM_TYPED_FUNC_START(sha256_transform_zvknha_or_zvknhb_zvkb)
|
|
|
|
// Load the round constants into K0-K15.
|
|
vsetivli zero, 4, e32, m1, ta, ma
|
|
la t0, K256
|
|
vle32.v K0, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K1, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K2, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K3, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K4, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K5, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K6, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K7, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K8, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K9, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K10, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K11, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K12, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K13, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K14, (t0)
|
|
addi t0, t0, 16
|
|
vle32.v K15, (t0)
|
|
|
|
// Setup mask for the vmerge to replace the first word (idx==0) in
|
|
// message scheduling. There are 4 words, so an 8-bit mask suffices.
|
|
vsetivli zero, 1, e8, m1, ta, ma
|
|
vmv.v.i MASK, 0x01
|
|
|
|
// Load the state. The state is stored as {a,b,c,d,e,f,g,h}, but we
|
|
// need {f,e,b,a},{h,g,d,c}. The dst vtype is e32m1 and the index vtype
|
|
// is e8mf4. We use index-load with the i8 indices {20, 16, 4, 0},
|
|
// loaded using the 32-bit little endian value 0x00041014.
|
|
li t0, 0x00041014
|
|
vsetivli zero, 1, e32, m1, ta, ma
|
|
vmv.v.x INDICES, t0
|
|
addi STATEP_C, STATEP, 8
|
|
vsetivli zero, 4, e32, m1, ta, ma
|
|
vluxei8.v FEBA, (STATEP), INDICES
|
|
vluxei8.v HGDC, (STATEP_C), INDICES
|
|
|
|
.Lnext_block:
|
|
addi NUM_BLOCKS, NUM_BLOCKS, -1
|
|
|
|
// Save the previous state, as it's needed later.
|
|
vmv.v.v PREV_FEBA, FEBA
|
|
vmv.v.v PREV_HGDC, HGDC
|
|
|
|
// Load the next 512-bit message block and endian-swap each 32-bit word.
|
|
vle32.v W0, (DATA)
|
|
vrev8.v W0, W0
|
|
addi DATA, DATA, 16
|
|
vle32.v W1, (DATA)
|
|
vrev8.v W1, W1
|
|
addi DATA, DATA, 16
|
|
vle32.v W2, (DATA)
|
|
vrev8.v W2, W2
|
|
addi DATA, DATA, 16
|
|
vle32.v W3, (DATA)
|
|
vrev8.v W3, W3
|
|
addi DATA, DATA, 16
|
|
|
|
// Do the 64 rounds of SHA-256.
|
|
sha256_16rounds 0, K0, K1, K2, K3
|
|
sha256_16rounds 0, K4, K5, K6, K7
|
|
sha256_16rounds 0, K8, K9, K10, K11
|
|
sha256_16rounds 1, K12, K13, K14, K15
|
|
|
|
// Add the previous state.
|
|
vadd.vv FEBA, FEBA, PREV_FEBA
|
|
vadd.vv HGDC, HGDC, PREV_HGDC
|
|
|
|
// Repeat if more blocks remain.
|
|
bnez NUM_BLOCKS, .Lnext_block
|
|
|
|
// Store the new state and return.
|
|
vsuxei8.v FEBA, (STATEP), INDICES
|
|
vsuxei8.v HGDC, (STATEP_C), INDICES
|
|
ret
|
|
SYM_FUNC_END(sha256_transform_zvknha_or_zvknhb_zvkb)
|
|
|
|
.section ".rodata"
|
|
.p2align 2
|
|
.type K256, @object
|
|
K256:
|
|
.word 0x428a2f98, 0x71374491, 0xb5c0fbcf, 0xe9b5dba5
|
|
.word 0x3956c25b, 0x59f111f1, 0x923f82a4, 0xab1c5ed5
|
|
.word 0xd807aa98, 0x12835b01, 0x243185be, 0x550c7dc3
|
|
.word 0x72be5d74, 0x80deb1fe, 0x9bdc06a7, 0xc19bf174
|
|
.word 0xe49b69c1, 0xefbe4786, 0x0fc19dc6, 0x240ca1cc
|
|
.word 0x2de92c6f, 0x4a7484aa, 0x5cb0a9dc, 0x76f988da
|
|
.word 0x983e5152, 0xa831c66d, 0xb00327c8, 0xbf597fc7
|
|
.word 0xc6e00bf3, 0xd5a79147, 0x06ca6351, 0x14292967
|
|
.word 0x27b70a85, 0x2e1b2138, 0x4d2c6dfc, 0x53380d13
|
|
.word 0x650a7354, 0x766a0abb, 0x81c2c92e, 0x92722c85
|
|
.word 0xa2bfe8a1, 0xa81a664b, 0xc24b8b70, 0xc76c51a3
|
|
.word 0xd192e819, 0xd6990624, 0xf40e3585, 0x106aa070
|
|
.word 0x19a4c116, 0x1e376c08, 0x2748774c, 0x34b0bcb5
|
|
.word 0x391c0cb3, 0x4ed8aa4a, 0x5b9cca4f, 0x682e6ff3
|
|
.word 0x748f82ee, 0x78a5636f, 0x84c87814, 0x8cc70208
|
|
.word 0x90befffa, 0xa4506ceb, 0xbef9a3f7, 0xc67178f2
|
|
.size K256, . - K256
|