00f463b38a
Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
84 lines
1.5 KiB
C
84 lines
1.5 KiB
C
/*
|
|
* Carry-less multiply operations.
|
|
* SPDX-License-Identifier: GPL-2.0-or-later
|
|
*
|
|
* Copyright (C) 2023 Linaro, Ltd.
|
|
*/
|
|
|
|
#ifndef CRYPTO_CLMUL_H
|
|
#define CRYPTO_CLMUL_H
|
|
|
|
#include "qemu/int128.h"
|
|
#include "host/crypto/clmul.h"
|
|
|
|
/**
|
|
* clmul_8x8_low:
|
|
*
|
|
* Perform eight 8x8->8 carry-less multiplies.
|
|
*/
|
|
uint64_t clmul_8x8_low(uint64_t, uint64_t);
|
|
|
|
/**
|
|
* clmul_8x4_even:
|
|
*
|
|
* Perform four 8x8->16 carry-less multiplies.
|
|
* The odd bytes of the inputs are ignored.
|
|
*/
|
|
uint64_t clmul_8x4_even(uint64_t, uint64_t);
|
|
|
|
/**
|
|
* clmul_8x4_odd:
|
|
*
|
|
* Perform four 8x8->16 carry-less multiplies.
|
|
* The even bytes of the inputs are ignored.
|
|
*/
|
|
uint64_t clmul_8x4_odd(uint64_t, uint64_t);
|
|
|
|
/**
|
|
* clmul_8x4_packed:
|
|
*
|
|
* Perform four 8x8->16 carry-less multiplies.
|
|
*/
|
|
uint64_t clmul_8x4_packed(uint32_t, uint32_t);
|
|
|
|
/**
|
|
* clmul_16x2_even:
|
|
*
|
|
* Perform two 16x16->32 carry-less multiplies.
|
|
* The odd words of the inputs are ignored.
|
|
*/
|
|
uint64_t clmul_16x2_even(uint64_t, uint64_t);
|
|
|
|
/**
|
|
* clmul_16x2_odd:
|
|
*
|
|
* Perform two 16x16->32 carry-less multiplies.
|
|
* The even words of the inputs are ignored.
|
|
*/
|
|
uint64_t clmul_16x2_odd(uint64_t, uint64_t);
|
|
|
|
/**
|
|
* clmul_32:
|
|
*
|
|
* Perform a 32x32->64 carry-less multiply.
|
|
*/
|
|
uint64_t clmul_32(uint32_t, uint32_t);
|
|
|
|
/**
|
|
* clmul_64:
|
|
*
|
|
* Perform a 64x64->128 carry-less multiply.
|
|
*/
|
|
Int128 clmul_64_gen(uint64_t, uint64_t);
|
|
|
|
static inline Int128 clmul_64(uint64_t a, uint64_t b)
|
|
{
|
|
if (HAVE_CLMUL_ACCEL) {
|
|
return clmul_64_accel(a, b);
|
|
} else {
|
|
return clmul_64_gen(a, b);
|
|
}
|
|
}
|
|
|
|
#endif /* CRYPTO_CLMUL_H */
|