| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Intel Keem Bay OCS HCU Crypto Driver. |
| * |
| * Copyright (C) 2018-2020 Intel Corporation |
| */ |
| |
| #include <linux/delay.h> |
| #include <linux/device.h> |
| #include <linux/iopoll.h> |
| #include <linux/irq.h> |
| #include <linux/module.h> |
| |
| #include <crypto/sha2.h> |
| |
| #include "ocs-hcu.h" |
| |
| /* Registers. */ |
| #define OCS_HCU_MODE 0x00 |
| #define OCS_HCU_CHAIN 0x04 |
| #define OCS_HCU_OPERATION 0x08 |
| #define OCS_HCU_KEY_0 0x0C |
| #define OCS_HCU_ISR 0x50 |
| #define OCS_HCU_IER 0x54 |
| #define OCS_HCU_STATUS 0x58 |
| #define OCS_HCU_MSG_LEN_LO 0x60 |
| #define OCS_HCU_MSG_LEN_HI 0x64 |
| #define OCS_HCU_KEY_BYTE_ORDER_CFG 0x80 |
| #define OCS_HCU_DMA_SRC_ADDR 0x400 |
| #define OCS_HCU_DMA_SRC_SIZE 0x408 |
| #define OCS_HCU_DMA_DST_SIZE 0x40C |
| #define OCS_HCU_DMA_DMA_MODE 0x410 |
| #define OCS_HCU_DMA_NEXT_SRC_DESCR 0x418 |
| #define OCS_HCU_DMA_MSI_ISR 0x480 |
| #define OCS_HCU_DMA_MSI_IER 0x484 |
| #define OCS_HCU_DMA_MSI_MASK 0x488 |
| |
| /* Register bit definitions. */ |
| #define HCU_MODE_ALGO_SHIFT 16 |
| #define HCU_MODE_HMAC_SHIFT 22 |
| |
| #define HCU_STATUS_BUSY BIT(0) |
| |
| #define HCU_BYTE_ORDER_SWAP BIT(0) |
| |
| #define HCU_IRQ_HASH_DONE BIT(2) |
| #define HCU_IRQ_HASH_ERR_MASK (BIT(3) | BIT(1) | BIT(0)) |
| |
| #define HCU_DMA_IRQ_SRC_DONE BIT(0) |
| #define HCU_DMA_IRQ_SAI_ERR BIT(2) |
| #define HCU_DMA_IRQ_BAD_COMP_ERR BIT(3) |
| #define HCU_DMA_IRQ_INBUF_RD_ERR BIT(4) |
| #define HCU_DMA_IRQ_INBUF_WD_ERR BIT(5) |
| #define HCU_DMA_IRQ_OUTBUF_WR_ERR BIT(6) |
| #define HCU_DMA_IRQ_OUTBUF_RD_ERR BIT(7) |
| #define HCU_DMA_IRQ_CRD_ERR BIT(8) |
| #define HCU_DMA_IRQ_ERR_MASK (HCU_DMA_IRQ_SAI_ERR | \ |
| HCU_DMA_IRQ_BAD_COMP_ERR | \ |
| HCU_DMA_IRQ_INBUF_RD_ERR | \ |
| HCU_DMA_IRQ_INBUF_WD_ERR | \ |
| HCU_DMA_IRQ_OUTBUF_WR_ERR | \ |
| HCU_DMA_IRQ_OUTBUF_RD_ERR | \ |
| HCU_DMA_IRQ_CRD_ERR) |
| |
| #define HCU_DMA_SNOOP_MASK (0x7 << 28) |
| #define HCU_DMA_SRC_LL_EN BIT(25) |
| #define HCU_DMA_EN BIT(31) |
| |
| #define OCS_HCU_ENDIANNESS_VALUE 0x2A |
| |
| #define HCU_DMA_MSI_UNMASK BIT(0) |
| #define HCU_DMA_MSI_DISABLE 0 |
| #define HCU_IRQ_DISABLE 0 |
| |
| #define OCS_HCU_START BIT(0) |
| #define OCS_HCU_TERMINATE BIT(1) |
| |
| #define OCS_LL_DMA_FLAG_TERMINATE BIT(31) |
| |
| #define OCS_HCU_HW_KEY_LEN_U32 (OCS_HCU_HW_KEY_LEN / sizeof(u32)) |
| |
| #define HCU_DATA_WRITE_ENDIANNESS_OFFSET 26 |
| |
| #define OCS_HCU_NUM_CHAINS_SHA256_224_SM3 (SHA256_DIGEST_SIZE / sizeof(u32)) |
| #define OCS_HCU_NUM_CHAINS_SHA384_512 (SHA512_DIGEST_SIZE / sizeof(u32)) |
| |
| /* |
| * While polling on a busy HCU, wait maximum 200us between one check and the |
| * other. |
| */ |
| #define OCS_HCU_WAIT_BUSY_RETRY_DELAY_US 200 |
| /* Wait on a busy HCU for maximum 1 second. */ |
| #define OCS_HCU_WAIT_BUSY_TIMEOUT_US 1000000 |
| |
| /** |
| * struct ocs_hcu_dma_entry - An entry in an OCS DMA linked list. |
| * @src_addr: Source address of the data. |
| * @src_len: Length of data to be fetched. |
| * @nxt_desc: Next descriptor to fetch. |
| * @ll_flags: Flags (Freeze @ terminate) for the DMA engine. |
| */ |
| struct ocs_hcu_dma_entry { |
| u32 src_addr; |
| u32 src_len; |
| u32 nxt_desc; |
| u32 ll_flags; |
| }; |
| |
| /** |
| * struct ocs_hcu_dma_list - OCS-specific DMA linked list. |
| * @head: The head of the list (points to the array backing the list). |
| * @tail: The current tail of the list; NULL if the list is empty. |
| * @dma_addr: The DMA address of @head (i.e., the DMA address of the backing |
| * array). |
| * @max_nents: Maximum number of entries in the list (i.e., number of elements |
| * in the backing array). |
| * |
| * The OCS DMA list is an array-backed list of OCS DMA descriptors. The array |
| * backing the list is allocated with dma_alloc_coherent() and pointed by |
| * @head. |
| */ |
| struct ocs_hcu_dma_list { |
| struct ocs_hcu_dma_entry *head; |
| struct ocs_hcu_dma_entry *tail; |
| dma_addr_t dma_addr; |
| size_t max_nents; |
| }; |
| |
| static inline u32 ocs_hcu_num_chains(enum ocs_hcu_algo algo) |
| { |
| switch (algo) { |
| case OCS_HCU_ALGO_SHA224: |
| case OCS_HCU_ALGO_SHA256: |
| case OCS_HCU_ALGO_SM3: |
| return OCS_HCU_NUM_CHAINS_SHA256_224_SM3; |
| case OCS_HCU_ALGO_SHA384: |
| case OCS_HCU_ALGO_SHA512: |
| return OCS_HCU_NUM_CHAINS_SHA384_512; |
| default: |
| return 0; |
| }; |
| } |
| |
| static inline u32 ocs_hcu_digest_size(enum ocs_hcu_algo algo) |
| { |
| switch (algo) { |
| case OCS_HCU_ALGO_SHA224: |
| return SHA224_DIGEST_SIZE; |
| case OCS_HCU_ALGO_SHA256: |
| case OCS_HCU_ALGO_SM3: |
| /* SM3 shares the same block size. */ |
| return SHA256_DIGEST_SIZE; |
| case OCS_HCU_ALGO_SHA384: |
| return SHA384_DIGEST_SIZE; |
| case OCS_HCU_ALGO_SHA512: |
| return SHA512_DIGEST_SIZE; |
| default: |
| return 0; |
| } |
| } |
| |
| /** |
| * ocs_hcu_wait_busy() - Wait for HCU OCS hardware to became usable. |
| * @hcu_dev: OCS HCU device to wait for. |
| * |
| * Return: 0 if device free, -ETIMEOUT if device busy and internal timeout has |
| * expired. |
| */ |
| static int ocs_hcu_wait_busy(struct ocs_hcu_dev *hcu_dev) |
| { |
| long val; |
| |
| return readl_poll_timeout(hcu_dev->io_base + OCS_HCU_STATUS, val, |
| !(val & HCU_STATUS_BUSY), |
| OCS_HCU_WAIT_BUSY_RETRY_DELAY_US, |
| OCS_HCU_WAIT_BUSY_TIMEOUT_US); |
| } |
| |
| static void ocs_hcu_done_irq_en(struct ocs_hcu_dev *hcu_dev) |
| { |
| /* Clear any pending interrupts. */ |
| writel(0xFFFFFFFF, hcu_dev->io_base + OCS_HCU_ISR); |
| hcu_dev->irq_err = false; |
| /* Enable error and HCU done interrupts. */ |
| writel(HCU_IRQ_HASH_DONE | HCU_IRQ_HASH_ERR_MASK, |
| hcu_dev->io_base + OCS_HCU_IER); |
| } |
| |
| static void ocs_hcu_dma_irq_en(struct ocs_hcu_dev *hcu_dev) |
| { |
| /* Clear any pending interrupts. */ |
| writel(0xFFFFFFFF, hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR); |
| hcu_dev->irq_err = false; |
| /* Only operating on DMA source completion and error interrupts. */ |
| writel(HCU_DMA_IRQ_ERR_MASK | HCU_DMA_IRQ_SRC_DONE, |
| hcu_dev->io_base + OCS_HCU_DMA_MSI_IER); |
| /* Unmask */ |
| writel(HCU_DMA_MSI_UNMASK, hcu_dev->io_base + OCS_HCU_DMA_MSI_MASK); |
| } |
| |
| static void ocs_hcu_irq_dis(struct ocs_hcu_dev *hcu_dev) |
| { |
| writel(HCU_IRQ_DISABLE, hcu_dev->io_base + OCS_HCU_IER); |
| writel(HCU_DMA_MSI_DISABLE, hcu_dev->io_base + OCS_HCU_DMA_MSI_IER); |
| } |
| |
| static int ocs_hcu_wait_and_disable_irq(struct ocs_hcu_dev *hcu_dev) |
| { |
| int rc; |
| |
| rc = wait_for_completion_interruptible(&hcu_dev->irq_done); |
| if (rc) |
| goto exit; |
| |
| if (hcu_dev->irq_err) { |
| /* Unset flag and return error. */ |
| hcu_dev->irq_err = false; |
| rc = -EIO; |
| goto exit; |
| } |
| |
| exit: |
| ocs_hcu_irq_dis(hcu_dev); |
| |
| return rc; |
| } |
| |
| /** |
| * ocs_hcu_get_intermediate_data() - Get intermediate data. |
| * @hcu_dev: The target HCU device. |
| * @data: Where to store the intermediate. |
| * @algo: The algorithm being used. |
| * |
| * This function is used to save the current hashing process state in order to |
| * continue it in the future. |
| * |
| * Note: once all data has been processed, the intermediate data actually |
| * contains the hashing result. So this function is also used to retrieve the |
| * final result of a hashing process. |
| * |
| * Return: 0 on success, negative error code otherwise. |
| */ |
| static int ocs_hcu_get_intermediate_data(struct ocs_hcu_dev *hcu_dev, |
| struct ocs_hcu_idata *data, |
| enum ocs_hcu_algo algo) |
| { |
| const int n = ocs_hcu_num_chains(algo); |
| u32 *chain; |
| int rc; |
| int i; |
| |
| /* Data not requested. */ |
| if (!data) |
| return -EINVAL; |
| |
| chain = (u32 *)data->digest; |
| |
| /* Ensure that the OCS is no longer busy before reading the chains. */ |
| rc = ocs_hcu_wait_busy(hcu_dev); |
| if (rc) |
| return rc; |
| |
| /* |
| * This loops is safe because data->digest is an array of |
| * SHA512_DIGEST_SIZE bytes and the maximum value returned by |
| * ocs_hcu_num_chains() is OCS_HCU_NUM_CHAINS_SHA384_512 which is equal |
| * to SHA512_DIGEST_SIZE / sizeof(u32). |
| */ |
| for (i = 0; i < n; i++) |
| chain[i] = readl(hcu_dev->io_base + OCS_HCU_CHAIN); |
| |
| data->msg_len_lo = readl(hcu_dev->io_base + OCS_HCU_MSG_LEN_LO); |
| data->msg_len_hi = readl(hcu_dev->io_base + OCS_HCU_MSG_LEN_HI); |
| |
| return 0; |
| } |
| |
| /** |
| * ocs_hcu_set_intermediate_data() - Set intermediate data. |
| * @hcu_dev: The target HCU device. |
| * @data: The intermediate data to be set. |
| * @algo: The algorithm being used. |
| * |
| * This function is used to continue a previous hashing process. |
| */ |
| static void ocs_hcu_set_intermediate_data(struct ocs_hcu_dev *hcu_dev, |
| const struct ocs_hcu_idata *data, |
| enum ocs_hcu_algo algo) |
| { |
| const int n = ocs_hcu_num_chains(algo); |
| u32 *chain = (u32 *)data->digest; |
| int i; |
| |
| /* |
| * This loops is safe because data->digest is an array of |
| * SHA512_DIGEST_SIZE bytes and the maximum value returned by |
| * ocs_hcu_num_chains() is OCS_HCU_NUM_CHAINS_SHA384_512 which is equal |
| * to SHA512_DIGEST_SIZE / sizeof(u32). |
| */ |
| for (i = 0; i < n; i++) |
| writel(chain[i], hcu_dev->io_base + OCS_HCU_CHAIN); |
| |
| writel(data->msg_len_lo, hcu_dev->io_base + OCS_HCU_MSG_LEN_LO); |
| writel(data->msg_len_hi, hcu_dev->io_base + OCS_HCU_MSG_LEN_HI); |
| } |
| |
| static int ocs_hcu_get_digest(struct ocs_hcu_dev *hcu_dev, |
| enum ocs_hcu_algo algo, u8 *dgst, size_t dgst_len) |
| { |
| u32 *chain; |
| int rc; |
| int i; |
| |
| if (!dgst) |
| return -EINVAL; |
| |
| /* Length of the output buffer must match the algo digest size. */ |
| if (dgst_len != ocs_hcu_digest_size(algo)) |
| return -EINVAL; |
| |
| /* Ensure that the OCS is no longer busy before reading the chains. */ |
| rc = ocs_hcu_wait_busy(hcu_dev); |
| if (rc) |
| return rc; |
| |
| chain = (u32 *)dgst; |
| for (i = 0; i < dgst_len / sizeof(u32); i++) |
| chain[i] = readl(hcu_dev->io_base + OCS_HCU_CHAIN); |
| |
| return 0; |
| } |
| |
| /** |
| * ocs_hcu_hw_cfg() - Configure the HCU hardware. |
| * @hcu_dev: The HCU device to configure. |
| * @algo: The algorithm to be used by the HCU device. |
| * @use_hmac: Whether or not HW HMAC should be used. |
| * |
| * Return: 0 on success, negative error code otherwise. |
| */ |
| static int ocs_hcu_hw_cfg(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo, |
| bool use_hmac) |
| { |
| u32 cfg; |
| int rc; |
| |
| if (algo != OCS_HCU_ALGO_SHA256 && algo != OCS_HCU_ALGO_SHA224 && |
| algo != OCS_HCU_ALGO_SHA384 && algo != OCS_HCU_ALGO_SHA512 && |
| algo != OCS_HCU_ALGO_SM3) |
| return -EINVAL; |
| |
| rc = ocs_hcu_wait_busy(hcu_dev); |
| if (rc) |
| return rc; |
| |
| /* Ensure interrupts are disabled. */ |
| ocs_hcu_irq_dis(hcu_dev); |
| |
| /* Configure endianness, hashing algorithm and HW HMAC (if needed) */ |
| cfg = OCS_HCU_ENDIANNESS_VALUE << HCU_DATA_WRITE_ENDIANNESS_OFFSET; |
| cfg |= algo << HCU_MODE_ALGO_SHIFT; |
| if (use_hmac) |
| cfg |= BIT(HCU_MODE_HMAC_SHIFT); |
| |
| writel(cfg, hcu_dev->io_base + OCS_HCU_MODE); |
| |
| return 0; |
| } |
| |
| /** |
| * ocs_hcu_clear_key() - Clear key stored in OCS HMAC KEY registers. |
| * @hcu_dev: The OCS HCU device whose key registers should be cleared. |
| */ |
| static void ocs_hcu_clear_key(struct ocs_hcu_dev *hcu_dev) |
| { |
| int reg_off; |
| |
| /* Clear OCS_HCU_KEY_[0..15] */ |
| for (reg_off = 0; reg_off < OCS_HCU_HW_KEY_LEN; reg_off += sizeof(u32)) |
| writel(0, hcu_dev->io_base + OCS_HCU_KEY_0 + reg_off); |
| } |
| |
| /** |
| * ocs_hcu_write_key() - Write key to OCS HMAC KEY registers. |
| * @hcu_dev: The OCS HCU device the key should be written to. |
| * @key: The key to be written. |
| * @len: The size of the key to write. It must be OCS_HCU_HW_KEY_LEN. |
| * |
| * Return: 0 on success, negative error code otherwise. |
| */ |
| static int ocs_hcu_write_key(struct ocs_hcu_dev *hcu_dev, const u8 *key, size_t len) |
| { |
| u32 key_u32[OCS_HCU_HW_KEY_LEN_U32]; |
| int i; |
| |
| if (len > OCS_HCU_HW_KEY_LEN) |
| return -EINVAL; |
| |
| /* Copy key into temporary u32 array. */ |
| memcpy(key_u32, key, len); |
| |
| /* |
| * Hardware requires all the bytes of the HW Key vector to be |
| * written. So pad with zero until we reach OCS_HCU_HW_KEY_LEN. |
| */ |
| memzero_explicit((u8 *)key_u32 + len, OCS_HCU_HW_KEY_LEN - len); |
| |
| /* |
| * OCS hardware expects the MSB of the key to be written at the highest |
| * address of the HCU Key vector; in other word, the key must be |
| * written in reverse order. |
| * |
| * Therefore, we first enable byte swapping for the HCU key vector; |
| * so that bytes of 32-bit word written to OCS_HCU_KEY_[0..15] will be |
| * swapped: |
| * 3 <---> 0, 2 <---> 1. |
| */ |
| writel(HCU_BYTE_ORDER_SWAP, |
| hcu_dev->io_base + OCS_HCU_KEY_BYTE_ORDER_CFG); |
| /* |
| * And then we write the 32-bit words composing the key starting from |
| * the end of the key. |
| */ |
| for (i = 0; i < OCS_HCU_HW_KEY_LEN_U32; i++) |
| writel(key_u32[OCS_HCU_HW_KEY_LEN_U32 - 1 - i], |
| hcu_dev->io_base + OCS_HCU_KEY_0 + (sizeof(u32) * i)); |
| |
| memzero_explicit(key_u32, OCS_HCU_HW_KEY_LEN); |
| |
| return 0; |
| } |
| |
| /** |
| * ocs_hcu_ll_dma_start() - Start OCS HCU hashing via DMA |
| * @hcu_dev: The OCS HCU device to use. |
| * @dma_list: The OCS DMA list mapping the data to hash. |
| * @finalize: Whether or not this is the last hashing operation and therefore |
| * the final hash should be compute even if data is not |
| * block-aligned. |
| * |
| * Return: 0 on success, negative error code otherwise. |
| */ |
| static int ocs_hcu_ll_dma_start(struct ocs_hcu_dev *hcu_dev, |
| const struct ocs_hcu_dma_list *dma_list, |
| bool finalize) |
| { |
| u32 cfg = HCU_DMA_SNOOP_MASK | HCU_DMA_SRC_LL_EN | HCU_DMA_EN; |
| int rc; |
| |
| if (!dma_list) |
| return -EINVAL; |
| |
| /* |
| * For final requests we use HCU_DONE IRQ to be notified when all input |
| * data has been processed by the HCU; however, we cannot do so for |
| * non-final requests, because we don't get a HCU_DONE IRQ when we |
| * don't terminate the operation. |
| * |
| * Therefore, for non-final requests, we use the DMA IRQ, which |
| * triggers when DMA has finishing feeding all the input data to the |
| * HCU, but the HCU may still be processing it. This is fine, since we |
| * will wait for the HCU processing to be completed when we try to read |
| * intermediate results, in ocs_hcu_get_intermediate_data(). |
| */ |
| if (finalize) |
| ocs_hcu_done_irq_en(hcu_dev); |
| else |
| ocs_hcu_dma_irq_en(hcu_dev); |
| |
| reinit_completion(&hcu_dev->irq_done); |
| writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR); |
| writel(0, hcu_dev->io_base + OCS_HCU_DMA_SRC_SIZE); |
| writel(0, hcu_dev->io_base + OCS_HCU_DMA_DST_SIZE); |
| |
| writel(OCS_HCU_START, hcu_dev->io_base + OCS_HCU_OPERATION); |
| |
| writel(cfg, hcu_dev->io_base + OCS_HCU_DMA_DMA_MODE); |
| |
| if (finalize) |
| writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION); |
| |
| rc = ocs_hcu_wait_and_disable_irq(hcu_dev); |
| if (rc) |
| return rc; |
| |
| return 0; |
| } |
| |
| struct ocs_hcu_dma_list *ocs_hcu_dma_list_alloc(struct ocs_hcu_dev *hcu_dev, |
| int max_nents) |
| { |
| struct ocs_hcu_dma_list *dma_list; |
| |
| dma_list = kmalloc(sizeof(*dma_list), GFP_KERNEL); |
| if (!dma_list) |
| return NULL; |
| |
| /* Total size of the DMA list to allocate. */ |
| dma_list->head = dma_alloc_coherent(hcu_dev->dev, |
| sizeof(*dma_list->head) * max_nents, |
| &dma_list->dma_addr, GFP_KERNEL); |
| if (!dma_list->head) { |
| kfree(dma_list); |
| return NULL; |
| } |
| dma_list->max_nents = max_nents; |
| dma_list->tail = NULL; |
| |
| return dma_list; |
| } |
| |
| void ocs_hcu_dma_list_free(struct ocs_hcu_dev *hcu_dev, |
| struct ocs_hcu_dma_list *dma_list) |
| { |
| if (!dma_list) |
| return; |
| |
| dma_free_coherent(hcu_dev->dev, |
| sizeof(*dma_list->head) * dma_list->max_nents, |
| dma_list->head, dma_list->dma_addr); |
| |
| kfree(dma_list); |
| } |
| |
| /* Add a new DMA entry at the end of the OCS DMA list. */ |
| int ocs_hcu_dma_list_add_tail(struct ocs_hcu_dev *hcu_dev, |
| struct ocs_hcu_dma_list *dma_list, |
| dma_addr_t addr, u32 len) |
| { |
| struct device *dev = hcu_dev->dev; |
| struct ocs_hcu_dma_entry *old_tail; |
| struct ocs_hcu_dma_entry *new_tail; |
| |
| if (!len) |
| return 0; |
| |
| if (!dma_list) |
| return -EINVAL; |
| |
| if (addr & ~OCS_HCU_DMA_BIT_MASK) { |
| dev_err(dev, |
| "Unexpected error: Invalid DMA address for OCS HCU\n"); |
| return -EINVAL; |
| } |
| |
| old_tail = dma_list->tail; |
| new_tail = old_tail ? old_tail + 1 : dma_list->head; |
| |
| /* Check if list is full. */ |
| if (new_tail - dma_list->head >= dma_list->max_nents) |
| return -ENOMEM; |
| |
| /* |
| * If there was an old tail (i.e., this is not the first element we are |
| * adding), un-terminate the old tail and make it point to the new one. |
| */ |
| if (old_tail) { |
| old_tail->ll_flags &= ~OCS_LL_DMA_FLAG_TERMINATE; |
| /* |
| * The old tail 'nxt_desc' must point to the DMA address of the |
| * new tail. |
| */ |
| old_tail->nxt_desc = dma_list->dma_addr + |
| sizeof(*dma_list->tail) * (new_tail - |
| dma_list->head); |
| } |
| |
| new_tail->src_addr = (u32)addr; |
| new_tail->src_len = (u32)len; |
| new_tail->ll_flags = OCS_LL_DMA_FLAG_TERMINATE; |
| new_tail->nxt_desc = 0; |
| |
| /* Update list tail with new tail. */ |
| dma_list->tail = new_tail; |
| |
| return 0; |
| } |
| |
| /** |
| * ocs_hcu_hash_init() - Initialize hash operation context. |
| * @ctx: The context to initialize. |
| * @algo: The hashing algorithm to use. |
| * |
| * Return: 0 on success, negative error code otherwise. |
| */ |
| int ocs_hcu_hash_init(struct ocs_hcu_hash_ctx *ctx, enum ocs_hcu_algo algo) |
| { |
| if (!ctx) |
| return -EINVAL; |
| |
| ctx->algo = algo; |
| ctx->idata.msg_len_lo = 0; |
| ctx->idata.msg_len_hi = 0; |
| /* No need to set idata.digest to 0. */ |
| |
| return 0; |
| } |
| |
| /** |
| * ocs_hcu_hash_update() - Perform a hashing iteration. |
| * @hcu_dev: The OCS HCU device to use. |
| * @ctx: The OCS HCU hashing context. |
| * @dma_list: The OCS DMA list mapping the input data to process. |
| * |
| * Return: 0 on success; negative error code otherwise. |
| */ |
| int ocs_hcu_hash_update(struct ocs_hcu_dev *hcu_dev, |
| struct ocs_hcu_hash_ctx *ctx, |
| const struct ocs_hcu_dma_list *dma_list) |
| { |
| int rc; |
| |
| if (!hcu_dev || !ctx) |
| return -EINVAL; |
| |
| /* Configure the hardware for the current request. */ |
| rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false); |
| if (rc) |
| return rc; |
| |
| /* If we already processed some data, idata needs to be set. */ |
| if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi) |
| ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo); |
| |
| /* Start linked-list DMA hashing. */ |
| rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, false); |
| if (rc) |
| return rc; |
| |
| /* Update idata and return. */ |
| return ocs_hcu_get_intermediate_data(hcu_dev, &ctx->idata, ctx->algo); |
| } |
| |
| /** |
| * ocs_hcu_hash_finup() - Update and finalize hash computation. |
| * @hcu_dev: The OCS HCU device to use. |
| * @ctx: The OCS HCU hashing context. |
| * @dma_list: The OCS DMA list mapping the input data to process. |
| * @dgst: The buffer where to save the computed digest. |
| * @dgst_len: The length of @dgst. |
| * |
| * Return: 0 on success; negative error code otherwise. |
| */ |
| int ocs_hcu_hash_finup(struct ocs_hcu_dev *hcu_dev, |
| const struct ocs_hcu_hash_ctx *ctx, |
| const struct ocs_hcu_dma_list *dma_list, |
| u8 *dgst, size_t dgst_len) |
| { |
| int rc; |
| |
| if (!hcu_dev || !ctx) |
| return -EINVAL; |
| |
| /* Configure the hardware for the current request. */ |
| rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false); |
| if (rc) |
| return rc; |
| |
| /* If we already processed some data, idata needs to be set. */ |
| if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi) |
| ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo); |
| |
| /* Start linked-list DMA hashing. */ |
| rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, true); |
| if (rc) |
| return rc; |
| |
| /* Get digest and return. */ |
| return ocs_hcu_get_digest(hcu_dev, ctx->algo, dgst, dgst_len); |
| } |
| |
| /** |
| * ocs_hcu_hash_final() - Finalize hash computation. |
| * @hcu_dev: The OCS HCU device to use. |
| * @ctx: The OCS HCU hashing context. |
| * @dgst: The buffer where to save the computed digest. |
| * @dgst_len: The length of @dgst. |
| * |
| * Return: 0 on success; negative error code otherwise. |
| */ |
| int ocs_hcu_hash_final(struct ocs_hcu_dev *hcu_dev, |
| const struct ocs_hcu_hash_ctx *ctx, u8 *dgst, |
| size_t dgst_len) |
| { |
| int rc; |
| |
| if (!hcu_dev || !ctx) |
| return -EINVAL; |
| |
| /* Configure the hardware for the current request. */ |
| rc = ocs_hcu_hw_cfg(hcu_dev, ctx->algo, false); |
| if (rc) |
| return rc; |
| |
| /* If we already processed some data, idata needs to be set. */ |
| if (ctx->idata.msg_len_lo || ctx->idata.msg_len_hi) |
| ocs_hcu_set_intermediate_data(hcu_dev, &ctx->idata, ctx->algo); |
| |
| /* |
| * Enable HCU interrupts, so that HCU_DONE will be triggered once the |
| * final hash is computed. |
| */ |
| ocs_hcu_done_irq_en(hcu_dev); |
| reinit_completion(&hcu_dev->irq_done); |
| writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION); |
| |
| rc = ocs_hcu_wait_and_disable_irq(hcu_dev); |
| if (rc) |
| return rc; |
| |
| /* Get digest and return. */ |
| return ocs_hcu_get_digest(hcu_dev, ctx->algo, dgst, dgst_len); |
| } |
| |
| /** |
| * ocs_hcu_digest() - Compute hash digest. |
| * @hcu_dev: The OCS HCU device to use. |
| * @algo: The hash algorithm to use. |
| * @data: The input data to process. |
| * @data_len: The length of @data. |
| * @dgst: The buffer where to save the computed digest. |
| * @dgst_len: The length of @dgst. |
| * |
| * Return: 0 on success; negative error code otherwise. |
| */ |
| int ocs_hcu_digest(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo, |
| void *data, size_t data_len, u8 *dgst, size_t dgst_len) |
| { |
| struct device *dev = hcu_dev->dev; |
| dma_addr_t dma_handle; |
| u32 reg; |
| int rc; |
| |
| /* Configure the hardware for the current request. */ |
| rc = ocs_hcu_hw_cfg(hcu_dev, algo, false); |
| if (rc) |
| return rc; |
| |
| dma_handle = dma_map_single(dev, data, data_len, DMA_TO_DEVICE); |
| if (dma_mapping_error(dev, dma_handle)) |
| return -EIO; |
| |
| reg = HCU_DMA_SNOOP_MASK | HCU_DMA_EN; |
| |
| ocs_hcu_done_irq_en(hcu_dev); |
| |
| reinit_completion(&hcu_dev->irq_done); |
| |
| writel(dma_handle, hcu_dev->io_base + OCS_HCU_DMA_SRC_ADDR); |
| writel(data_len, hcu_dev->io_base + OCS_HCU_DMA_SRC_SIZE); |
| writel(OCS_HCU_START, hcu_dev->io_base + OCS_HCU_OPERATION); |
| writel(reg, hcu_dev->io_base + OCS_HCU_DMA_DMA_MODE); |
| |
| writel(OCS_HCU_TERMINATE, hcu_dev->io_base + OCS_HCU_OPERATION); |
| |
| rc = ocs_hcu_wait_and_disable_irq(hcu_dev); |
| if (rc) |
| return rc; |
| |
| dma_unmap_single(dev, dma_handle, data_len, DMA_TO_DEVICE); |
| |
| return ocs_hcu_get_digest(hcu_dev, algo, dgst, dgst_len); |
| } |
| |
| /** |
| * ocs_hcu_hmac() - Compute HMAC. |
| * @hcu_dev: The OCS HCU device to use. |
| * @algo: The hash algorithm to use with HMAC. |
| * @key: The key to use. |
| * @dma_list: The OCS DMA list mapping the input data to process. |
| * @key_len: The length of @key. |
| * @dgst: The buffer where to save the computed HMAC. |
| * @dgst_len: The length of @dgst. |
| * |
| * Return: 0 on success; negative error code otherwise. |
| */ |
| int ocs_hcu_hmac(struct ocs_hcu_dev *hcu_dev, enum ocs_hcu_algo algo, |
| const u8 *key, size_t key_len, |
| const struct ocs_hcu_dma_list *dma_list, |
| u8 *dgst, size_t dgst_len) |
| { |
| int rc; |
| |
| /* Ensure 'key' is not NULL. */ |
| if (!key || key_len == 0) |
| return -EINVAL; |
| |
| /* Configure the hardware for the current request. */ |
| rc = ocs_hcu_hw_cfg(hcu_dev, algo, true); |
| if (rc) |
| return rc; |
| |
| rc = ocs_hcu_write_key(hcu_dev, key, key_len); |
| if (rc) |
| return rc; |
| |
| rc = ocs_hcu_ll_dma_start(hcu_dev, dma_list, true); |
| |
| /* Clear HW key before processing return code. */ |
| ocs_hcu_clear_key(hcu_dev); |
| |
| if (rc) |
| return rc; |
| |
| return ocs_hcu_get_digest(hcu_dev, algo, dgst, dgst_len); |
| } |
| |
| irqreturn_t ocs_hcu_irq_handler(int irq, void *dev_id) |
| { |
| struct ocs_hcu_dev *hcu_dev = dev_id; |
| u32 hcu_irq; |
| u32 dma_irq; |
| |
| /* Read and clear the HCU interrupt. */ |
| hcu_irq = readl(hcu_dev->io_base + OCS_HCU_ISR); |
| writel(hcu_irq, hcu_dev->io_base + OCS_HCU_ISR); |
| |
| /* Read and clear the HCU DMA interrupt. */ |
| dma_irq = readl(hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR); |
| writel(dma_irq, hcu_dev->io_base + OCS_HCU_DMA_MSI_ISR); |
| |
| /* Check for errors. */ |
| if (hcu_irq & HCU_IRQ_HASH_ERR_MASK || dma_irq & HCU_DMA_IRQ_ERR_MASK) { |
| hcu_dev->irq_err = true; |
| goto complete; |
| } |
| |
| /* Check for DONE IRQs. */ |
| if (hcu_irq & HCU_IRQ_HASH_DONE || dma_irq & HCU_DMA_IRQ_SRC_DONE) |
| goto complete; |
| |
| return IRQ_NONE; |
| |
| complete: |
| complete(&hcu_dev->irq_done); |
| |
| return IRQ_HANDLED; |
| } |
| |
| MODULE_DESCRIPTION("Intel Keem Bay OCS HCU Crypto Driver"); |
| MODULE_LICENSE("GPL"); |