| /* SPDX-License-Identifier: GPL-2.0 */ |
| /* Copyright (c) 2016-2017 Hisilicon Limited. */ |
| |
| #ifndef _SEC_DRV_H_ |
| #define _SEC_DRV_H_ |
| |
| #include <crypto/algapi.h> |
| #include <linux/kfifo.h> |
| |
| #define SEC_MAX_SGE_NUM 64 |
| #define SEC_HW_RING_NUM 3 |
| |
| #define SEC_CMD_RING 0 |
| #define SEC_OUTORDER_RING 1 |
| #define SEC_DBG_RING 2 |
| |
| /* A reasonable length to balance memory use against flexibility */ |
| #define SEC_QUEUE_LEN 512 |
| |
| #define SEC_MAX_SGE_NUM 64 |
| |
| struct sec_bd_info { |
| #define SEC_BD_W0_T_LEN_M GENMASK(4, 0) |
| #define SEC_BD_W0_T_LEN_S 0 |
| |
| #define SEC_BD_W0_C_WIDTH_M GENMASK(6, 5) |
| #define SEC_BD_W0_C_WIDTH_S 5 |
| #define SEC_C_WIDTH_AES_128BIT 0 |
| #define SEC_C_WIDTH_AES_8BIT 1 |
| #define SEC_C_WIDTH_AES_1BIT 2 |
| #define SEC_C_WIDTH_DES_64BIT 0 |
| #define SEC_C_WIDTH_DES_8BIT 1 |
| #define SEC_C_WIDTH_DES_1BIT 2 |
| |
| #define SEC_BD_W0_C_MODE_M GENMASK(9, 7) |
| #define SEC_BD_W0_C_MODE_S 7 |
| #define SEC_C_MODE_ECB 0 |
| #define SEC_C_MODE_CBC 1 |
| #define SEC_C_MODE_CTR 4 |
| #define SEC_C_MODE_CCM 5 |
| #define SEC_C_MODE_GCM 6 |
| #define SEC_C_MODE_XTS 7 |
| |
| #define SEC_BD_W0_SEQ BIT(10) |
| #define SEC_BD_W0_DE BIT(11) |
| #define SEC_BD_W0_DAT_SKIP_M GENMASK(13, 12) |
| #define SEC_BD_W0_DAT_SKIP_S 12 |
| #define SEC_BD_W0_C_GRAN_SIZE_19_16_M GENMASK(17, 14) |
| #define SEC_BD_W0_C_GRAN_SIZE_19_16_S 14 |
| |
| #define SEC_BD_W0_CIPHER_M GENMASK(19, 18) |
| #define SEC_BD_W0_CIPHER_S 18 |
| #define SEC_CIPHER_NULL 0 |
| #define SEC_CIPHER_ENCRYPT 1 |
| #define SEC_CIPHER_DECRYPT 2 |
| |
| #define SEC_BD_W0_AUTH_M GENMASK(21, 20) |
| #define SEC_BD_W0_AUTH_S 20 |
| #define SEC_AUTH_NULL 0 |
| #define SEC_AUTH_MAC 1 |
| #define SEC_AUTH_VERIF 2 |
| |
| #define SEC_BD_W0_AI_GEN BIT(22) |
| #define SEC_BD_W0_CI_GEN BIT(23) |
| #define SEC_BD_W0_NO_HPAD BIT(24) |
| #define SEC_BD_W0_HM_M GENMASK(26, 25) |
| #define SEC_BD_W0_HM_S 25 |
| #define SEC_BD_W0_ICV_OR_SKEY_EN_M GENMASK(28, 27) |
| #define SEC_BD_W0_ICV_OR_SKEY_EN_S 27 |
| |
| /* Multi purpose field - gran size bits for send, flag for recv */ |
| #define SEC_BD_W0_FLAG_M GENMASK(30, 29) |
| #define SEC_BD_W0_C_GRAN_SIZE_21_20_M GENMASK(30, 29) |
| #define SEC_BD_W0_FLAG_S 29 |
| #define SEC_BD_W0_C_GRAN_SIZE_21_20_S 29 |
| |
| #define SEC_BD_W0_DONE BIT(31) |
| u32 w0; |
| |
| #define SEC_BD_W1_AUTH_GRAN_SIZE_M GENMASK(21, 0) |
| #define SEC_BD_W1_AUTH_GRAN_SIZE_S 0 |
| #define SEC_BD_W1_M_KEY_EN BIT(22) |
| #define SEC_BD_W1_BD_INVALID BIT(23) |
| #define SEC_BD_W1_ADDR_TYPE BIT(24) |
| |
| #define SEC_BD_W1_A_ALG_M GENMASK(28, 25) |
| #define SEC_BD_W1_A_ALG_S 25 |
| #define SEC_A_ALG_SHA1 0 |
| #define SEC_A_ALG_SHA256 1 |
| #define SEC_A_ALG_MD5 2 |
| #define SEC_A_ALG_SHA224 3 |
| #define SEC_A_ALG_HMAC_SHA1 8 |
| #define SEC_A_ALG_HMAC_SHA224 10 |
| #define SEC_A_ALG_HMAC_SHA256 11 |
| #define SEC_A_ALG_HMAC_MD5 12 |
| #define SEC_A_ALG_AES_XCBC 13 |
| #define SEC_A_ALG_AES_CMAC 14 |
| |
| #define SEC_BD_W1_C_ALG_M GENMASK(31, 29) |
| #define SEC_BD_W1_C_ALG_S 29 |
| #define SEC_C_ALG_DES 0 |
| #define SEC_C_ALG_3DES 1 |
| #define SEC_C_ALG_AES 2 |
| |
| u32 w1; |
| |
| #define SEC_BD_W2_C_GRAN_SIZE_15_0_M GENMASK(15, 0) |
| #define SEC_BD_W2_C_GRAN_SIZE_15_0_S 0 |
| #define SEC_BD_W2_GRAN_NUM_M GENMASK(31, 16) |
| #define SEC_BD_W2_GRAN_NUM_S 16 |
| u32 w2; |
| |
| #define SEC_BD_W3_AUTH_LEN_OFFSET_M GENMASK(9, 0) |
| #define SEC_BD_W3_AUTH_LEN_OFFSET_S 0 |
| #define SEC_BD_W3_CIPHER_LEN_OFFSET_M GENMASK(19, 10) |
| #define SEC_BD_W3_CIPHER_LEN_OFFSET_S 10 |
| #define SEC_BD_W3_MAC_LEN_M GENMASK(24, 20) |
| #define SEC_BD_W3_MAC_LEN_S 20 |
| #define SEC_BD_W3_A_KEY_LEN_M GENMASK(29, 25) |
| #define SEC_BD_W3_A_KEY_LEN_S 25 |
| #define SEC_BD_W3_C_KEY_LEN_M GENMASK(31, 30) |
| #define SEC_BD_W3_C_KEY_LEN_S 30 |
| #define SEC_KEY_LEN_AES_128 0 |
| #define SEC_KEY_LEN_AES_192 1 |
| #define SEC_KEY_LEN_AES_256 2 |
| #define SEC_KEY_LEN_DES 1 |
| #define SEC_KEY_LEN_3DES_3_KEY 1 |
| #define SEC_KEY_LEN_3DES_2_KEY 3 |
| u32 w3; |
| |
| /* W4,5 */ |
| union { |
| u32 authkey_addr_lo; |
| u32 authiv_addr_lo; |
| }; |
| union { |
| u32 authkey_addr_hi; |
| u32 authiv_addr_hi; |
| }; |
| |
| /* W6,7 */ |
| u32 cipher_key_addr_lo; |
| u32 cipher_key_addr_hi; |
| |
| /* W8,9 */ |
| u32 cipher_iv_addr_lo; |
| u32 cipher_iv_addr_hi; |
| |
| /* W10,11 */ |
| u32 data_addr_lo; |
| u32 data_addr_hi; |
| |
| /* W12,13 */ |
| u32 mac_addr_lo; |
| u32 mac_addr_hi; |
| |
| /* W14,15 */ |
| u32 cipher_destin_addr_lo; |
| u32 cipher_destin_addr_hi; |
| }; |
| |
| enum sec_mem_region { |
| SEC_COMMON = 0, |
| SEC_SAA, |
| SEC_NUM_ADDR_REGIONS |
| }; |
| |
| #define SEC_NAME_SIZE 64 |
| #define SEC_Q_NUM 16 |
| |
| |
| /** |
| * struct sec_queue_ring_cmd - store information about a SEC HW cmd ring |
| * @used: Local counter used to cheaply establish if the ring is empty. |
| * @lock: Protect against simultaneous adjusting of the read and write pointers. |
| * @vaddr: Virtual address for the ram pages used for the ring. |
| * @paddr: Physical address of the dma mapped region of ram used for the ring. |
| * @callback: Callback function called on a ring element completing. |
| */ |
| struct sec_queue_ring_cmd { |
| atomic_t used; |
| struct mutex lock; |
| struct sec_bd_info *vaddr; |
| dma_addr_t paddr; |
| void (*callback)(struct sec_bd_info *resp, void *ctx); |
| }; |
| |
| struct sec_debug_bd_info; |
| struct sec_queue_ring_db { |
| struct sec_debug_bd_info *vaddr; |
| dma_addr_t paddr; |
| }; |
| |
| struct sec_out_bd_info; |
| struct sec_queue_ring_cq { |
| struct sec_out_bd_info *vaddr; |
| dma_addr_t paddr; |
| }; |
| |
| struct sec_dev_info; |
| |
| enum sec_cipher_alg { |
| SEC_C_DES_ECB_64, |
| SEC_C_DES_CBC_64, |
| |
| SEC_C_3DES_ECB_192_3KEY, |
| SEC_C_3DES_ECB_192_2KEY, |
| |
| SEC_C_3DES_CBC_192_3KEY, |
| SEC_C_3DES_CBC_192_2KEY, |
| |
| SEC_C_AES_ECB_128, |
| SEC_C_AES_ECB_192, |
| SEC_C_AES_ECB_256, |
| |
| SEC_C_AES_CBC_128, |
| SEC_C_AES_CBC_192, |
| SEC_C_AES_CBC_256, |
| |
| SEC_C_AES_CTR_128, |
| SEC_C_AES_CTR_192, |
| SEC_C_AES_CTR_256, |
| |
| SEC_C_AES_XTS_128, |
| SEC_C_AES_XTS_256, |
| |
| SEC_C_NULL, |
| }; |
| |
| /** |
| * struct sec_alg_tfm_ctx - hardware specific tranformation context |
| * @cipher_alg: Cipher algorithm enabled include encryption mode. |
| * @key: Key storage if required. |
| * @pkey: DMA address for the key storage. |
| * @req_template: Request template to save time on setup. |
| * @queue: The hardware queue associated with this tfm context. |
| * @lock: Protect key and pkey to ensure they are consistent |
| * @auth_buf: Current context buffer for auth operations. |
| * @backlog: The backlog queue used for cases where our buffers aren't |
| * large enough. |
| */ |
| struct sec_alg_tfm_ctx { |
| enum sec_cipher_alg cipher_alg; |
| u8 *key; |
| dma_addr_t pkey; |
| struct sec_bd_info req_template; |
| struct sec_queue *queue; |
| struct mutex lock; |
| u8 *auth_buf; |
| struct list_head backlog; |
| }; |
| |
| /** |
| * struct sec_request - data associate with a single crypto request |
| * @elements: List of subparts of this request (hardware size restriction) |
| * @num_elements: The number of subparts (used as an optimization) |
| * @lock: Protect elements of this structure against concurrent change. |
| * @tfm_ctx: hardware specific context. |
| * @len_in: length of in sgl from upper layers |
| * @len_out: length of out sgl from upper layers |
| * @dma_iv: initialization vector - phsyical address |
| * @err: store used to track errors across subelements of this request. |
| * @req_base: pointer to base element of associate crypto context. |
| * This is needed to allow shared handling skcipher, ahash etc. |
| * @cb: completion callback. |
| * @backlog_head: list head to allow backlog maintenance. |
| * |
| * The hardware is limited in the maximum size of data that it can |
| * process from a single BD. Typically this is fairly large (32MB) |
| * but still requires the complexity of splitting the incoming |
| * skreq up into a number of elements complete with appropriate |
| * iv chaining. |
| */ |
| struct sec_request { |
| struct list_head elements; |
| int num_elements; |
| struct mutex lock; |
| struct sec_alg_tfm_ctx *tfm_ctx; |
| int len_in; |
| int len_out; |
| dma_addr_t dma_iv; |
| int err; |
| struct crypto_async_request *req_base; |
| void (*cb)(struct sec_bd_info *resp, struct crypto_async_request *req); |
| struct list_head backlog_head; |
| }; |
| |
| /** |
| * struct sec_request_el - A subpart of a request. |
| * @head: allow us to attach this to the list in the sec_request |
| * @req: hardware block descriptor corresponding to this request subpart |
| * @in: hardware sgl for input - virtual address |
| * @dma_in: hardware sgl for input - physical address |
| * @sgl_in: scatterlist for this request subpart |
| * @out: hardware sgl for output - virtual address |
| * @dma_out: hardware sgl for output - physical address |
| * @sgl_out: scatterlist for this request subpart |
| * @sec_req: The request which this subpart forms a part of |
| * @el_length: Number of bytes in this subpart. Needed to locate |
| * last ivsize chunk for iv chaining. |
| */ |
| struct sec_request_el { |
| struct list_head head; |
| struct sec_bd_info req; |
| struct sec_hw_sgl *in; |
| dma_addr_t dma_in; |
| struct scatterlist *sgl_in; |
| struct sec_hw_sgl *out; |
| dma_addr_t dma_out; |
| struct scatterlist *sgl_out; |
| struct sec_request *sec_req; |
| size_t el_length; |
| }; |
| |
| /** |
| * struct sec_queue - All the information about a HW queue |
| * @dev_info: The parent SEC device to which this queue belongs. |
| * @task_irq: Completion interrupt for the queue. |
| * @name: Human readable queue description also used as irq name. |
| * @ring: The several HW rings associated with one queue. |
| * @regs: The iomapped device registers |
| * @queue_id: Index of the queue used for naming and resource selection. |
| * @in_use: Flag to say if the queue is in use. |
| * @expected: The next expected element to finish assuming we were in order. |
| * @uprocessed: A bitmap to track which OoO elements are done but not handled. |
| * @softqueue: A software queue used when chaining requirements prevent direct |
| * use of the hardware queues. |
| * @havesoftqueue: A flag to say we have a queues - as we may need one for the |
| * current mode. |
| * @queuelock: Protect the soft queue from concurrent changes to avoid some |
| * potential loss of data races. |
| * @shadow: Pointers back to the shadow copy of the hardware ring element |
| * need because we can't store any context reference in the bd element. |
| */ |
| struct sec_queue { |
| struct sec_dev_info *dev_info; |
| int task_irq; |
| char name[SEC_NAME_SIZE]; |
| struct sec_queue_ring_cmd ring_cmd; |
| struct sec_queue_ring_cq ring_cq; |
| struct sec_queue_ring_db ring_db; |
| void __iomem *regs; |
| u32 queue_id; |
| bool in_use; |
| int expected; |
| |
| DECLARE_BITMAP(unprocessed, SEC_QUEUE_LEN); |
| DECLARE_KFIFO_PTR(softqueue, typeof(struct sec_request_el *)); |
| bool havesoftqueue; |
| struct mutex queuelock; |
| void *shadow[SEC_QUEUE_LEN]; |
| }; |
| |
| /** |
| * struct sec_hw_sge: Track each of the 64 element SEC HW SGL entries |
| * @buf: The IOV dma address for this entry. |
| * @len: Length of this IOV. |
| * @pad: Reserved space. |
| */ |
| struct sec_hw_sge { |
| dma_addr_t buf; |
| unsigned int len; |
| unsigned int pad; |
| }; |
| |
| /** |
| * struct sec_hw_sgl: One hardware SGL entry. |
| * @next_sgl: The next entry if we need to chain dma address. Null if last. |
| * @entry_sum_in_chain: The full count of SGEs - only matters for first SGL. |
| * @entry_sum_in_sgl: The number of SGEs in this SGL element. |
| * @flag: Unused in skciphers. |
| * @serial_num: Unsued in skciphers. |
| * @cpuid: Currently unused. |
| * @data_bytes_in_sgl: Count of bytes from all SGEs in this SGL. |
| * @next: Virtual address used to stash the next sgl - useful in completion. |
| * @reserved: A reserved field not currently used. |
| * @sge_entries: The (up to) 64 Scatter Gather Entries, representing IOVs. |
| * @node: Currently unused. |
| */ |
| struct sec_hw_sgl { |
| dma_addr_t next_sgl; |
| u16 entry_sum_in_chain; |
| u16 entry_sum_in_sgl; |
| u32 flag; |
| u64 serial_num; |
| u32 cpuid; |
| u32 data_bytes_in_sgl; |
| struct sec_hw_sgl *next; |
| u64 reserved; |
| struct sec_hw_sge sge_entries[SEC_MAX_SGE_NUM]; |
| u8 node[16]; |
| }; |
| |
| struct dma_pool; |
| |
| /** |
| * struct sec_dev_info: The full SEC unit comprising queues and processors. |
| * @sec_id: Index used to track which SEC this is when more than one is present. |
| * @num_saas: The number of backed processors enabled. |
| * @regs: iomapped register regions shared by whole SEC unit. |
| * @dev_lock: Protects concurrent queue allocation / freeing for the SEC. |
| * @queues: The 16 queues that this SEC instance provides. |
| * @dev: Device pointer. |
| * @hw_sgl_pool: DMA pool used to mimise mapping for the scatter gather lists. |
| */ |
| struct sec_dev_info { |
| int sec_id; |
| int num_saas; |
| void __iomem *regs[SEC_NUM_ADDR_REGIONS]; |
| struct mutex dev_lock; |
| int queues_in_use; |
| struct sec_queue queues[SEC_Q_NUM]; |
| struct device *dev; |
| struct dma_pool *hw_sgl_pool; |
| }; |
| |
| int sec_queue_send(struct sec_queue *queue, struct sec_bd_info *msg, void *ctx); |
| bool sec_queue_can_enqueue(struct sec_queue *queue, int num); |
| int sec_queue_stop_release(struct sec_queue *queue); |
| struct sec_queue *sec_queue_alloc_start_safe(void); |
| bool sec_queue_empty(struct sec_queue *queue); |
| |
| /* Algorithm specific elements from sec_algs.c */ |
| void sec_alg_callback(struct sec_bd_info *resp, void *ctx); |
| int sec_algs_register(void); |
| void sec_algs_unregister(void); |
| |
| #endif /* _SEC_DRV_H_ */ |