| // SPDX-License-Identifier: GPL-2.0-or-later |
| /* |
| * Cryptographic API. |
| * |
| * Single-block cipher operations. |
| * |
| * Copyright (c) 2002 James Morris <jmorris@intercode.com.au> |
| * Copyright (c) 2005 Herbert Xu <herbert@gondor.apana.org.au> |
| */ |
| |
| #include <crypto/algapi.h> |
| #include <crypto/internal/cipher.h> |
| #include <linux/kernel.h> |
| #include <linux/crypto.h> |
| #include <linux/errno.h> |
| #include <linux/slab.h> |
| #include <linux/string.h> |
| #include "internal.h" |
| |
| static int setkey_unaligned(struct crypto_cipher *tfm, const u8 *key, |
| unsigned int keylen) |
| { |
| struct cipher_alg *cia = crypto_cipher_alg(tfm); |
| unsigned long alignmask = crypto_cipher_alignmask(tfm); |
| int ret; |
| u8 *buffer, *alignbuffer; |
| unsigned long absize; |
| |
| absize = keylen + alignmask; |
| buffer = kmalloc(absize, GFP_ATOMIC); |
| if (!buffer) |
| return -ENOMEM; |
| |
| alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); |
| memcpy(alignbuffer, key, keylen); |
| ret = cia->cia_setkey(crypto_cipher_tfm(tfm), alignbuffer, keylen); |
| kfree_sensitive(buffer); |
| return ret; |
| |
| } |
| |
| int crypto_cipher_setkey(struct crypto_cipher *tfm, |
| const u8 *key, unsigned int keylen) |
| { |
| struct cipher_alg *cia = crypto_cipher_alg(tfm); |
| unsigned long alignmask = crypto_cipher_alignmask(tfm); |
| |
| if (keylen < cia->cia_min_keysize || keylen > cia->cia_max_keysize) |
| return -EINVAL; |
| |
| if ((unsigned long)key & alignmask) |
| return setkey_unaligned(tfm, key, keylen); |
| |
| return cia->cia_setkey(crypto_cipher_tfm(tfm), key, keylen); |
| } |
| EXPORT_SYMBOL_NS_GPL(crypto_cipher_setkey, "CRYPTO_INTERNAL"); |
| |
| static inline void cipher_crypt_one(struct crypto_cipher *tfm, |
| u8 *dst, const u8 *src, bool enc) |
| { |
| unsigned long alignmask = crypto_cipher_alignmask(tfm); |
| struct cipher_alg *cia = crypto_cipher_alg(tfm); |
| void (*fn)(struct crypto_tfm *, u8 *, const u8 *) = |
| enc ? cia->cia_encrypt : cia->cia_decrypt; |
| |
| if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { |
| unsigned int bs = crypto_cipher_blocksize(tfm); |
| u8 buffer[MAX_CIPHER_BLOCKSIZE + MAX_CIPHER_ALIGNMASK]; |
| u8 *tmp = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); |
| |
| memcpy(tmp, src, bs); |
| fn(crypto_cipher_tfm(tfm), tmp, tmp); |
| memcpy(dst, tmp, bs); |
| } else { |
| fn(crypto_cipher_tfm(tfm), dst, src); |
| } |
| } |
| |
| void crypto_cipher_encrypt_one(struct crypto_cipher *tfm, |
| u8 *dst, const u8 *src) |
| { |
| cipher_crypt_one(tfm, dst, src, true); |
| } |
| EXPORT_SYMBOL_NS_GPL(crypto_cipher_encrypt_one, "CRYPTO_INTERNAL"); |
| |
| void crypto_cipher_decrypt_one(struct crypto_cipher *tfm, |
| u8 *dst, const u8 *src) |
| { |
| cipher_crypt_one(tfm, dst, src, false); |
| } |
| EXPORT_SYMBOL_NS_GPL(crypto_cipher_decrypt_one, "CRYPTO_INTERNAL"); |
| |
| struct crypto_cipher *crypto_clone_cipher(struct crypto_cipher *cipher) |
| { |
| struct crypto_tfm *tfm = crypto_cipher_tfm(cipher); |
| struct crypto_alg *alg = tfm->__crt_alg; |
| struct crypto_cipher *ncipher; |
| struct crypto_tfm *ntfm; |
| |
| if (alg->cra_init) |
| return ERR_PTR(-ENOSYS); |
| |
| if (unlikely(!crypto_mod_get(alg))) |
| return ERR_PTR(-ESTALE); |
| |
| ntfm = __crypto_alloc_tfmgfp(alg, CRYPTO_ALG_TYPE_CIPHER, |
| CRYPTO_ALG_TYPE_MASK, GFP_ATOMIC); |
| if (IS_ERR(ntfm)) { |
| crypto_mod_put(alg); |
| return ERR_CAST(ntfm); |
| } |
| |
| ntfm->crt_flags = tfm->crt_flags; |
| |
| ncipher = __crypto_cipher_cast(ntfm); |
| |
| return ncipher; |
| } |
| EXPORT_SYMBOL_GPL(crypto_clone_cipher); |