| // SPDX-License-Identifier: GPL-2.0-or-later |
| /* |
| * CBC: Cipher Block Chaining mode |
| * |
| * Copyright (c) 2006-2016 Herbert Xu <herbert@gondor.apana.org.au> |
| */ |
| |
| #include <crypto/algapi.h> |
| #include <crypto/internal/cipher.h> |
| #include <crypto/internal/skcipher.h> |
| #include <linux/err.h> |
| #include <linux/init.h> |
| #include <linux/kernel.h> |
| #include <linux/log2.h> |
| #include <linux/module.h> |
| |
| static int crypto_cbc_encrypt_segment(struct skcipher_walk *walk, |
| struct crypto_skcipher *skcipher) |
| { |
| unsigned int bsize = crypto_skcipher_blocksize(skcipher); |
| void (*fn)(struct crypto_tfm *, u8 *, const u8 *); |
| unsigned int nbytes = walk->nbytes; |
| u8 *src = walk->src.virt.addr; |
| u8 *dst = walk->dst.virt.addr; |
| struct crypto_cipher *cipher; |
| struct crypto_tfm *tfm; |
| u8 *iv = walk->iv; |
| |
| cipher = skcipher_cipher_simple(skcipher); |
| tfm = crypto_cipher_tfm(cipher); |
| fn = crypto_cipher_alg(cipher)->cia_encrypt; |
| |
| do { |
| crypto_xor(iv, src, bsize); |
| fn(tfm, dst, iv); |
| memcpy(iv, dst, bsize); |
| |
| src += bsize; |
| dst += bsize; |
| } while ((nbytes -= bsize) >= bsize); |
| |
| return nbytes; |
| } |
| |
| static int crypto_cbc_encrypt_inplace(struct skcipher_walk *walk, |
| struct crypto_skcipher *skcipher) |
| { |
| unsigned int bsize = crypto_skcipher_blocksize(skcipher); |
| void (*fn)(struct crypto_tfm *, u8 *, const u8 *); |
| unsigned int nbytes = walk->nbytes; |
| u8 *src = walk->src.virt.addr; |
| struct crypto_cipher *cipher; |
| struct crypto_tfm *tfm; |
| u8 *iv = walk->iv; |
| |
| cipher = skcipher_cipher_simple(skcipher); |
| tfm = crypto_cipher_tfm(cipher); |
| fn = crypto_cipher_alg(cipher)->cia_encrypt; |
| |
| do { |
| crypto_xor(src, iv, bsize); |
| fn(tfm, src, src); |
| iv = src; |
| |
| src += bsize; |
| } while ((nbytes -= bsize) >= bsize); |
| |
| memcpy(walk->iv, iv, bsize); |
| |
| return nbytes; |
| } |
| |
| static int crypto_cbc_encrypt(struct skcipher_request *req) |
| { |
| struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); |
| struct skcipher_walk walk; |
| int err; |
| |
| err = skcipher_walk_virt(&walk, req, false); |
| |
| while (walk.nbytes) { |
| if (walk.src.virt.addr == walk.dst.virt.addr) |
| err = crypto_cbc_encrypt_inplace(&walk, skcipher); |
| else |
| err = crypto_cbc_encrypt_segment(&walk, skcipher); |
| err = skcipher_walk_done(&walk, err); |
| } |
| |
| return err; |
| } |
| |
| static int crypto_cbc_decrypt_segment(struct skcipher_walk *walk, |
| struct crypto_skcipher *skcipher) |
| { |
| unsigned int bsize = crypto_skcipher_blocksize(skcipher); |
| void (*fn)(struct crypto_tfm *, u8 *, const u8 *); |
| unsigned int nbytes = walk->nbytes; |
| u8 *src = walk->src.virt.addr; |
| u8 *dst = walk->dst.virt.addr; |
| struct crypto_cipher *cipher; |
| struct crypto_tfm *tfm; |
| u8 *iv = walk->iv; |
| |
| cipher = skcipher_cipher_simple(skcipher); |
| tfm = crypto_cipher_tfm(cipher); |
| fn = crypto_cipher_alg(cipher)->cia_decrypt; |
| |
| do { |
| fn(tfm, dst, src); |
| crypto_xor(dst, iv, bsize); |
| iv = src; |
| |
| src += bsize; |
| dst += bsize; |
| } while ((nbytes -= bsize) >= bsize); |
| |
| memcpy(walk->iv, iv, bsize); |
| |
| return nbytes; |
| } |
| |
| static int crypto_cbc_decrypt_inplace(struct skcipher_walk *walk, |
| struct crypto_skcipher *skcipher) |
| { |
| unsigned int bsize = crypto_skcipher_blocksize(skcipher); |
| void (*fn)(struct crypto_tfm *, u8 *, const u8 *); |
| unsigned int nbytes = walk->nbytes; |
| u8 *src = walk->src.virt.addr; |
| u8 last_iv[MAX_CIPHER_BLOCKSIZE]; |
| struct crypto_cipher *cipher; |
| struct crypto_tfm *tfm; |
| |
| cipher = skcipher_cipher_simple(skcipher); |
| tfm = crypto_cipher_tfm(cipher); |
| fn = crypto_cipher_alg(cipher)->cia_decrypt; |
| |
| /* Start of the last block. */ |
| src += nbytes - (nbytes & (bsize - 1)) - bsize; |
| memcpy(last_iv, src, bsize); |
| |
| for (;;) { |
| fn(tfm, src, src); |
| if ((nbytes -= bsize) < bsize) |
| break; |
| crypto_xor(src, src - bsize, bsize); |
| src -= bsize; |
| } |
| |
| crypto_xor(src, walk->iv, bsize); |
| memcpy(walk->iv, last_iv, bsize); |
| |
| return nbytes; |
| } |
| |
| static int crypto_cbc_decrypt(struct skcipher_request *req) |
| { |
| struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); |
| struct skcipher_walk walk; |
| int err; |
| |
| err = skcipher_walk_virt(&walk, req, false); |
| |
| while (walk.nbytes) { |
| if (walk.src.virt.addr == walk.dst.virt.addr) |
| err = crypto_cbc_decrypt_inplace(&walk, skcipher); |
| else |
| err = crypto_cbc_decrypt_segment(&walk, skcipher); |
| err = skcipher_walk_done(&walk, err); |
| } |
| |
| return err; |
| } |
| |
| static int crypto_cbc_create(struct crypto_template *tmpl, struct rtattr **tb) |
| { |
| struct skcipher_instance *inst; |
| struct crypto_alg *alg; |
| int err; |
| |
| inst = skcipher_alloc_instance_simple(tmpl, tb); |
| if (IS_ERR(inst)) |
| return PTR_ERR(inst); |
| |
| alg = skcipher_ialg_simple(inst); |
| |
| err = -EINVAL; |
| if (!is_power_of_2(alg->cra_blocksize)) |
| goto out_free_inst; |
| |
| inst->alg.encrypt = crypto_cbc_encrypt; |
| inst->alg.decrypt = crypto_cbc_decrypt; |
| |
| err = skcipher_register_instance(tmpl, inst); |
| if (err) { |
| out_free_inst: |
| inst->free(inst); |
| } |
| |
| return err; |
| } |
| |
| static struct crypto_template crypto_cbc_tmpl = { |
| .name = "cbc", |
| .create = crypto_cbc_create, |
| .module = THIS_MODULE, |
| }; |
| |
| static int __init crypto_cbc_module_init(void) |
| { |
| return crypto_register_template(&crypto_cbc_tmpl); |
| } |
| |
| static void __exit crypto_cbc_module_exit(void) |
| { |
| crypto_unregister_template(&crypto_cbc_tmpl); |
| } |
| |
| subsys_initcall(crypto_cbc_module_init); |
| module_exit(crypto_cbc_module_exit); |
| |
| MODULE_LICENSE("GPL"); |
| MODULE_DESCRIPTION("CBC block cipher mode of operation"); |
| MODULE_ALIAS_CRYPTO("cbc"); |