| // SPDX-License-Identifier: GPL-2.0-only | 
 | /* | 
 |  * Intel Keem Bay OCS HCU Crypto Driver. | 
 |  * | 
 |  * Copyright (C) 2018-2020 Intel Corporation | 
 |  */ | 
 |  | 
 | #include <crypto/engine.h> | 
 | #include <crypto/hmac.h> | 
 | #include <crypto/internal/hash.h> | 
 | #include <crypto/scatterwalk.h> | 
 | #include <crypto/sha2.h> | 
 | #include <crypto/sm3.h> | 
 | #include <linux/completion.h> | 
 | #include <linux/dma-mapping.h> | 
 | #include <linux/err.h> | 
 | #include <linux/interrupt.h> | 
 | #include <linux/kernel.h> | 
 | #include <linux/mod_devicetable.h> | 
 | #include <linux/module.h> | 
 | #include <linux/platform_device.h> | 
 | #include <linux/string.h> | 
 |  | 
 | #include "ocs-hcu.h" | 
 |  | 
 | #define DRV_NAME	"keembay-ocs-hcu" | 
 |  | 
 | /* Flag marking a final request. */ | 
 | #define REQ_FINAL			BIT(0) | 
 | /* Flag marking a HMAC request. */ | 
 | #define REQ_FLAGS_HMAC			BIT(1) | 
 | /* Flag set when HW HMAC is being used. */ | 
 | #define REQ_FLAGS_HMAC_HW		BIT(2) | 
 | /* Flag set when SW HMAC is being used. */ | 
 | #define REQ_FLAGS_HMAC_SW		BIT(3) | 
 |  | 
 | /** | 
 |  * struct ocs_hcu_ctx: OCS HCU Transform context. | 
 |  * @hcu_dev:	 The OCS HCU device used by the transformation. | 
 |  * @key:	 The key (used only for HMAC transformations). | 
 |  * @key_len:	 The length of the key. | 
 |  * @is_sm3_tfm:  Whether or not this is an SM3 transformation. | 
 |  * @is_hmac_tfm: Whether or not this is a HMAC transformation. | 
 |  */ | 
 | struct ocs_hcu_ctx { | 
 | 	struct ocs_hcu_dev *hcu_dev; | 
 | 	u8 key[SHA512_BLOCK_SIZE]; | 
 | 	size_t key_len; | 
 | 	bool is_sm3_tfm; | 
 | 	bool is_hmac_tfm; | 
 | }; | 
 |  | 
 | /** | 
 |  * struct ocs_hcu_rctx - Context for the request. | 
 |  * @hcu_dev:	    OCS HCU device to be used to service the request. | 
 |  * @flags:	    Flags tracking request status. | 
 |  * @algo:	    Algorithm to use for the request. | 
 |  * @blk_sz:	    Block size of the transformation / request. | 
 |  * @dig_sz:	    Digest size of the transformation / request. | 
 |  * @dma_list:	    OCS DMA linked list. | 
 |  * @hash_ctx:	    OCS HCU hashing context. | 
 |  * @buffer:	    Buffer to store: partial block of data and SW HMAC | 
 |  *		    artifacts (ipad, opad, etc.). | 
 |  * @buf_cnt:	    Number of bytes currently stored in the buffer. | 
 |  * @buf_dma_addr:   The DMA address of @buffer (when mapped). | 
 |  * @buf_dma_count:  The number of bytes in @buffer currently DMA-mapped. | 
 |  * @sg:		    Head of the scatterlist entries containing data. | 
 |  * @sg_data_total:  Total data in the SG list at any time. | 
 |  * @sg_data_offset: Offset into the data of the current individual SG node. | 
 |  * @sg_dma_nents:   Number of sg entries mapped in dma_list. | 
 |  * @nents:          Number of entries in the scatterlist. | 
 |  */ | 
 | struct ocs_hcu_rctx { | 
 | 	struct ocs_hcu_dev	*hcu_dev; | 
 | 	u32			flags; | 
 | 	enum ocs_hcu_algo	algo; | 
 | 	size_t			blk_sz; | 
 | 	size_t			dig_sz; | 
 | 	struct ocs_hcu_dma_list	*dma_list; | 
 | 	struct ocs_hcu_hash_ctx	hash_ctx; | 
 | 	/* | 
 | 	 * Buffer is double the block size because we need space for SW HMAC | 
 | 	 * artifacts, i.e: | 
 | 	 * - ipad (1 block) + a possible partial block of data. | 
 | 	 * - opad (1 block) + digest of H(k ^ ipad || m) | 
 | 	 */ | 
 | 	u8			buffer[2 * SHA512_BLOCK_SIZE]; | 
 | 	size_t			buf_cnt; | 
 | 	dma_addr_t		buf_dma_addr; | 
 | 	size_t			buf_dma_count; | 
 | 	struct scatterlist	*sg; | 
 | 	unsigned int		sg_data_total; | 
 | 	unsigned int		sg_data_offset; | 
 | 	unsigned int		sg_dma_nents; | 
 | 	unsigned int		nents; | 
 | }; | 
 |  | 
 | /** | 
 |  * struct ocs_hcu_drv - Driver data | 
 |  * @dev_list:	The list of HCU devices. | 
 |  * @lock:	The lock protecting dev_list. | 
 |  */ | 
 | struct ocs_hcu_drv { | 
 | 	struct list_head dev_list; | 
 | 	spinlock_t lock; /* Protects dev_list. */ | 
 | }; | 
 |  | 
 | static struct ocs_hcu_drv ocs_hcu = { | 
 | 	.dev_list = LIST_HEAD_INIT(ocs_hcu.dev_list), | 
 | 	.lock = __SPIN_LOCK_UNLOCKED(ocs_hcu.lock), | 
 | }; | 
 |  | 
 | /* | 
 |  * Return the total amount of data in the request; that is: the data in the | 
 |  * request buffer + the data in the sg list. | 
 |  */ | 
 | static inline unsigned int kmb_get_total_data(struct ocs_hcu_rctx *rctx) | 
 | { | 
 | 	return rctx->sg_data_total + rctx->buf_cnt; | 
 | } | 
 |  | 
 | /* Move remaining content of scatter-gather list to context buffer. */ | 
 | static int flush_sg_to_ocs_buffer(struct ocs_hcu_rctx *rctx) | 
 | { | 
 | 	size_t count; | 
 |  | 
 | 	if (rctx->sg_data_total > (sizeof(rctx->buffer) - rctx->buf_cnt)) { | 
 | 		WARN(1, "%s: sg data does not fit in buffer\n", __func__); | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	while (rctx->sg_data_total) { | 
 | 		if (!rctx->sg) { | 
 | 			WARN(1, "%s: unexpected NULL sg\n", __func__); | 
 | 			return -EINVAL; | 
 | 		} | 
 | 		/* | 
 | 		 * If current sg has been fully processed, skip to the next | 
 | 		 * one. | 
 | 		 */ | 
 | 		if (rctx->sg_data_offset == rctx->sg->length) { | 
 | 			rctx->sg = sg_next(rctx->sg); | 
 | 			rctx->sg_data_offset = 0; | 
 | 			continue; | 
 | 		} | 
 | 		/* | 
 | 		 * Determine the maximum data available to copy from the node. | 
 | 		 * Minimum of the length left in the sg node, or the total data | 
 | 		 * in the request. | 
 | 		 */ | 
 | 		count = min(rctx->sg->length - rctx->sg_data_offset, | 
 | 			    rctx->sg_data_total); | 
 | 		/* Copy from scatter-list entry to context buffer. */ | 
 | 		scatterwalk_map_and_copy(&rctx->buffer[rctx->buf_cnt], | 
 | 					 rctx->sg, rctx->sg_data_offset, | 
 | 					 count, 0); | 
 |  | 
 | 		rctx->sg_data_offset += count; | 
 | 		rctx->sg_data_total -= count; | 
 | 		rctx->buf_cnt += count; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static struct ocs_hcu_dev *kmb_ocs_hcu_find_dev(struct ahash_request *req) | 
 | { | 
 | 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); | 
 | 	struct ocs_hcu_ctx *tctx = crypto_ahash_ctx(tfm); | 
 |  | 
 | 	/* If the HCU device for the request was previously set, return it. */ | 
 | 	if (tctx->hcu_dev) | 
 | 		return tctx->hcu_dev; | 
 |  | 
 | 	/* | 
 | 	 * Otherwise, get the first HCU device available (there should be one | 
 | 	 * and only one device). | 
 | 	 */ | 
 | 	spin_lock_bh(&ocs_hcu.lock); | 
 | 	tctx->hcu_dev = list_first_entry_or_null(&ocs_hcu.dev_list, | 
 | 						 struct ocs_hcu_dev, | 
 | 						 list); | 
 | 	spin_unlock_bh(&ocs_hcu.lock); | 
 |  | 
 | 	return tctx->hcu_dev; | 
 | } | 
 |  | 
 | /* Free OCS DMA linked list and DMA-able context buffer. */ | 
 | static void kmb_ocs_hcu_dma_cleanup(struct ahash_request *req, | 
 | 				    struct ocs_hcu_rctx *rctx) | 
 | { | 
 | 	struct ocs_hcu_dev *hcu_dev = rctx->hcu_dev; | 
 | 	struct device *dev = hcu_dev->dev; | 
 |  | 
 | 	/* Unmap rctx->buffer (if mapped). */ | 
 | 	if (rctx->buf_dma_count) { | 
 | 		dma_unmap_single(dev, rctx->buf_dma_addr, rctx->buf_dma_count, | 
 | 				 DMA_TO_DEVICE); | 
 | 		rctx->buf_dma_count = 0; | 
 | 	} | 
 |  | 
 | 	/* Unmap req->src (if mapped). */ | 
 | 	if (rctx->sg_dma_nents) { | 
 | 		dma_unmap_sg(dev, req->src, rctx->nents, DMA_TO_DEVICE); | 
 | 		rctx->sg_dma_nents = 0; | 
 | 	} | 
 |  | 
 | 	/* Free dma_list (if allocated). */ | 
 | 	if (rctx->dma_list) { | 
 | 		ocs_hcu_dma_list_free(hcu_dev, rctx->dma_list); | 
 | 		rctx->dma_list = NULL; | 
 | 	} | 
 | } | 
 |  | 
 | /* | 
 |  * Prepare for DMA operation: | 
 |  * - DMA-map request context buffer (if needed) | 
 |  * - DMA-map SG list (only the entries to be processed, see note below) | 
 |  * - Allocate OCS HCU DMA linked list (number of elements =  SG entries to | 
 |  *   process + context buffer (if not empty)). | 
 |  * - Add DMA-mapped request context buffer to OCS HCU DMA list. | 
 |  * - Add SG entries to DMA list. | 
 |  * | 
 |  * Note: if this is a final request, we process all the data in the SG list, | 
 |  * otherwise we can only process up to the maximum amount of block-aligned data | 
 |  * (the remainder will be put into the context buffer and processed in the next | 
 |  * request). | 
 |  */ | 
 | static int kmb_ocs_dma_prepare(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 | 	struct device *dev = rctx->hcu_dev->dev; | 
 | 	unsigned int remainder = 0; | 
 | 	unsigned int total; | 
 | 	size_t nents; | 
 | 	size_t count; | 
 | 	int rc; | 
 | 	int i; | 
 |  | 
 | 	/* This function should be called only when there is data to process. */ | 
 | 	total = kmb_get_total_data(rctx); | 
 | 	if (!total) | 
 | 		return -EINVAL; | 
 |  | 
 | 	/* | 
 | 	 * If this is not a final DMA (terminated DMA), the data passed to the | 
 | 	 * HCU must be aligned to the block size; compute the remainder data to | 
 | 	 * be processed in the next request. | 
 | 	 */ | 
 | 	if (!(rctx->flags & REQ_FINAL)) | 
 | 		remainder = total % rctx->blk_sz; | 
 |  | 
 | 	/* Determine the number of scatter gather list entries to process. */ | 
 | 	nents = sg_nents_for_len(req->src, rctx->sg_data_total - remainder); | 
 |  | 
 | 	/* If there are entries to process, map them. */ | 
 | 	if (nents) { | 
 | 		rctx->sg_dma_nents = dma_map_sg(dev, req->src, nents, | 
 | 						DMA_TO_DEVICE); | 
 | 		if (!rctx->sg_dma_nents) { | 
 | 			dev_err(dev, "Failed to MAP SG\n"); | 
 | 			rc = -ENOMEM; | 
 | 			goto cleanup; | 
 | 		} | 
 |  | 
 | 		/* Save the value of nents to pass to dma_unmap_sg. */ | 
 | 		rctx->nents = nents; | 
 |  | 
 | 		/* | 
 | 		 * The value returned by dma_map_sg() can be < nents; so update | 
 | 		 * nents accordingly. | 
 | 		 */ | 
 | 		nents = rctx->sg_dma_nents; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * If context buffer is not empty, map it and add extra DMA entry for | 
 | 	 * it. | 
 | 	 */ | 
 | 	if (rctx->buf_cnt) { | 
 | 		rctx->buf_dma_addr = dma_map_single(dev, rctx->buffer, | 
 | 						    rctx->buf_cnt, | 
 | 						    DMA_TO_DEVICE); | 
 | 		if (dma_mapping_error(dev, rctx->buf_dma_addr)) { | 
 | 			dev_err(dev, "Failed to map request context buffer\n"); | 
 | 			rc = -ENOMEM; | 
 | 			goto cleanup; | 
 | 		} | 
 | 		rctx->buf_dma_count = rctx->buf_cnt; | 
 | 		/* Increase number of dma entries. */ | 
 | 		nents++; | 
 | 	} | 
 |  | 
 | 	/* Allocate OCS HCU DMA list. */ | 
 | 	rctx->dma_list = ocs_hcu_dma_list_alloc(rctx->hcu_dev, nents); | 
 | 	if (!rctx->dma_list) { | 
 | 		rc = -ENOMEM; | 
 | 		goto cleanup; | 
 | 	} | 
 |  | 
 | 	/* Add request context buffer (if previously DMA-mapped) */ | 
 | 	if (rctx->buf_dma_count) { | 
 | 		rc = ocs_hcu_dma_list_add_tail(rctx->hcu_dev, rctx->dma_list, | 
 | 					       rctx->buf_dma_addr, | 
 | 					       rctx->buf_dma_count); | 
 | 		if (rc) | 
 | 			goto cleanup; | 
 | 	} | 
 |  | 
 | 	/* Add the SG nodes to be processed to the DMA linked list. */ | 
 | 	for_each_sg(req->src, rctx->sg, rctx->sg_dma_nents, i) { | 
 | 		/* | 
 | 		 * The number of bytes to add to the list entry is the minimum | 
 | 		 * between: | 
 | 		 * - The DMA length of the SG entry. | 
 | 		 * - The data left to be processed. | 
 | 		 */ | 
 | 		count = min(rctx->sg_data_total - remainder, | 
 | 			    sg_dma_len(rctx->sg) - rctx->sg_data_offset); | 
 | 		/* | 
 | 		 * Do not create a zero length DMA descriptor. Check in case of | 
 | 		 * zero length SG node. | 
 | 		 */ | 
 | 		if (count == 0) | 
 | 			continue; | 
 | 		/* Add sg to HCU DMA list. */ | 
 | 		rc = ocs_hcu_dma_list_add_tail(rctx->hcu_dev, | 
 | 					       rctx->dma_list, | 
 | 					       rctx->sg->dma_address, | 
 | 					       count); | 
 | 		if (rc) | 
 | 			goto cleanup; | 
 |  | 
 | 		/* Update amount of data remaining in SG list. */ | 
 | 		rctx->sg_data_total -= count; | 
 |  | 
 | 		/* | 
 | 		 * If  remaining data is equal to remainder (note: 'less than' | 
 | 		 * case should never happen in practice), we are done: update | 
 | 		 * offset and exit the loop. | 
 | 		 */ | 
 | 		if (rctx->sg_data_total <= remainder) { | 
 | 			WARN_ON(rctx->sg_data_total < remainder); | 
 | 			rctx->sg_data_offset += count; | 
 | 			break; | 
 | 		} | 
 |  | 
 | 		/* | 
 | 		 * If we get here is because we need to process the next sg in | 
 | 		 * the list; set offset within the sg to 0. | 
 | 		 */ | 
 | 		rctx->sg_data_offset = 0; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | cleanup: | 
 | 	dev_err(dev, "Failed to prepare DMA.\n"); | 
 | 	kmb_ocs_hcu_dma_cleanup(req, rctx); | 
 |  | 
 | 	return rc; | 
 | } | 
 |  | 
 | static void kmb_ocs_hcu_secure_cleanup(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 |  | 
 | 	/* Clear buffer of any data. */ | 
 | 	memzero_explicit(rctx->buffer, sizeof(rctx->buffer)); | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_handle_queue(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_dev *hcu_dev = kmb_ocs_hcu_find_dev(req); | 
 |  | 
 | 	if (!hcu_dev) | 
 | 		return -ENOENT; | 
 |  | 
 | 	return crypto_transfer_hash_request_to_engine(hcu_dev->engine, req); | 
 | } | 
 |  | 
 | static int prepare_ipad(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 | 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); | 
 | 	struct ocs_hcu_ctx *ctx = crypto_ahash_ctx(tfm); | 
 | 	int i; | 
 |  | 
 | 	WARN(rctx->buf_cnt, "%s: Context buffer is not empty\n", __func__); | 
 | 	WARN(!(rctx->flags & REQ_FLAGS_HMAC_SW), | 
 | 	     "%s: HMAC_SW flag is not set\n", __func__); | 
 | 	/* | 
 | 	 * Key length must be equal to block size. If key is shorter, | 
 | 	 * we pad it with zero (note: key cannot be longer, since | 
 | 	 * longer keys are hashed by kmb_ocs_hcu_setkey()). | 
 | 	 */ | 
 | 	if (ctx->key_len > rctx->blk_sz) { | 
 | 		WARN(1, "%s: Invalid key length in tfm context\n", __func__); | 
 | 		return -EINVAL; | 
 | 	} | 
 | 	memzero_explicit(&ctx->key[ctx->key_len], | 
 | 			 rctx->blk_sz - ctx->key_len); | 
 | 	ctx->key_len = rctx->blk_sz; | 
 | 	/* | 
 | 	 * Prepare IPAD for HMAC. Only done for first block. | 
 | 	 * HMAC(k,m) = H(k ^ opad || H(k ^ ipad || m)) | 
 | 	 * k ^ ipad will be first hashed block. | 
 | 	 * k ^ opad will be calculated in the final request. | 
 | 	 * Only needed if not using HW HMAC. | 
 | 	 */ | 
 | 	for (i = 0; i < rctx->blk_sz; i++) | 
 | 		rctx->buffer[i] = ctx->key[i] ^ HMAC_IPAD_VALUE; | 
 | 	rctx->buf_cnt = rctx->blk_sz; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_do_one_request(struct crypto_engine *engine, void *areq) | 
 | { | 
 | 	struct ahash_request *req = container_of(areq, struct ahash_request, | 
 | 						 base); | 
 | 	struct ocs_hcu_dev *hcu_dev = kmb_ocs_hcu_find_dev(req); | 
 | 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 | 	struct ocs_hcu_ctx *tctx = crypto_ahash_ctx(tfm); | 
 | 	int rc; | 
 | 	int i; | 
 |  | 
 | 	if (!hcu_dev) { | 
 | 		rc = -ENOENT; | 
 | 		goto error; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * If hardware HMAC flag is set, perform HMAC in hardware. | 
 | 	 * | 
 | 	 * NOTE: this flag implies REQ_FINAL && kmb_get_total_data(rctx) | 
 | 	 */ | 
 | 	if (rctx->flags & REQ_FLAGS_HMAC_HW) { | 
 | 		/* Map input data into the HCU DMA linked list. */ | 
 | 		rc = kmb_ocs_dma_prepare(req); | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 		rc = ocs_hcu_hmac(hcu_dev, rctx->algo, tctx->key, tctx->key_len, | 
 | 				  rctx->dma_list, req->result, rctx->dig_sz); | 
 |  | 
 | 		/* Unmap data and free DMA list regardless of return code. */ | 
 | 		kmb_ocs_hcu_dma_cleanup(req, rctx); | 
 |  | 
 | 		/* Process previous return code. */ | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 		goto done; | 
 | 	} | 
 |  | 
 | 	/* Handle update request case. */ | 
 | 	if (!(rctx->flags & REQ_FINAL)) { | 
 | 		/* Update should always have input data. */ | 
 | 		if (!kmb_get_total_data(rctx)) | 
 | 			return -EINVAL; | 
 |  | 
 | 		/* Map input data into the HCU DMA linked list. */ | 
 | 		rc = kmb_ocs_dma_prepare(req); | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 		/* Do hashing step. */ | 
 | 		rc = ocs_hcu_hash_update(hcu_dev, &rctx->hash_ctx, | 
 | 					 rctx->dma_list); | 
 |  | 
 | 		/* Unmap data and free DMA list regardless of return code. */ | 
 | 		kmb_ocs_hcu_dma_cleanup(req, rctx); | 
 |  | 
 | 		/* Process previous return code. */ | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 		/* | 
 | 		 * Reset request buffer count (data in the buffer was just | 
 | 		 * processed). | 
 | 		 */ | 
 | 		rctx->buf_cnt = 0; | 
 | 		/* | 
 | 		 * Move remaining sg data into the request buffer, so that it | 
 | 		 * will be processed during the next request. | 
 | 		 * | 
 | 		 * NOTE: we have remaining data if kmb_get_total_data() was not | 
 | 		 * a multiple of block size. | 
 | 		 */ | 
 | 		rc = flush_sg_to_ocs_buffer(rctx); | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 		goto done; | 
 | 	} | 
 |  | 
 | 	/* If we get here, this is a final request. */ | 
 |  | 
 | 	/* If there is data to process, use finup. */ | 
 | 	if (kmb_get_total_data(rctx)) { | 
 | 		/* Map input data into the HCU DMA linked list. */ | 
 | 		rc = kmb_ocs_dma_prepare(req); | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 		/* Do hashing step. */ | 
 | 		rc = ocs_hcu_hash_finup(hcu_dev, &rctx->hash_ctx, | 
 | 					rctx->dma_list, | 
 | 					req->result, rctx->dig_sz); | 
 | 		/* Free DMA list regardless of return code. */ | 
 | 		kmb_ocs_hcu_dma_cleanup(req, rctx); | 
 |  | 
 | 		/* Process previous return code. */ | 
 | 		if (rc) | 
 | 			goto error; | 
 |  | 
 | 	} else {  /* Otherwise (if we have no data), use final. */ | 
 | 		rc = ocs_hcu_hash_final(hcu_dev, &rctx->hash_ctx, req->result, | 
 | 					rctx->dig_sz); | 
 | 		if (rc) | 
 | 			goto error; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * If we are finalizing a SW HMAC request, we just computed the result | 
 | 	 * of: H(k ^ ipad || m). | 
 | 	 * | 
 | 	 * We now need to complete the HMAC calculation with the OPAD step, | 
 | 	 * that is, we need to compute H(k ^ opad || digest), where digest is | 
 | 	 * the digest we just obtained, i.e., H(k ^ ipad || m). | 
 | 	 */ | 
 | 	if (rctx->flags & REQ_FLAGS_HMAC_SW) { | 
 | 		/* | 
 | 		 * Compute k ^ opad and store it in the request buffer (which | 
 | 		 * is not used anymore at this point). | 
 | 		 * Note: key has been padded / hashed already (so keylen == | 
 | 		 * blksz) . | 
 | 		 */ | 
 | 		WARN_ON(tctx->key_len != rctx->blk_sz); | 
 | 		for (i = 0; i < rctx->blk_sz; i++) | 
 | 			rctx->buffer[i] = tctx->key[i] ^ HMAC_OPAD_VALUE; | 
 | 		/* Now append the digest to the rest of the buffer. */ | 
 | 		for (i = 0; (i < rctx->dig_sz); i++) | 
 | 			rctx->buffer[rctx->blk_sz + i] = req->result[i]; | 
 |  | 
 | 		/* Now hash the buffer to obtain the final HMAC. */ | 
 | 		rc = ocs_hcu_digest(hcu_dev, rctx->algo, rctx->buffer, | 
 | 				    rctx->blk_sz + rctx->dig_sz, req->result, | 
 | 				    rctx->dig_sz); | 
 | 		if (rc) | 
 | 			goto error; | 
 | 	} | 
 |  | 
 | 	/* Perform secure clean-up. */ | 
 | 	kmb_ocs_hcu_secure_cleanup(req); | 
 | done: | 
 | 	crypto_finalize_hash_request(hcu_dev->engine, req, 0); | 
 |  | 
 | 	return 0; | 
 |  | 
 | error: | 
 | 	kmb_ocs_hcu_secure_cleanup(req); | 
 | 	return rc; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_init(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_dev *hcu_dev = kmb_ocs_hcu_find_dev(req); | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 | 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); | 
 | 	struct ocs_hcu_ctx *ctx = crypto_ahash_ctx(tfm); | 
 |  | 
 | 	if (!hcu_dev) | 
 | 		return -ENOENT; | 
 |  | 
 | 	/* Initialize entire request context to zero. */ | 
 | 	memset(rctx, 0, sizeof(*rctx)); | 
 |  | 
 | 	rctx->hcu_dev = hcu_dev; | 
 | 	rctx->dig_sz = crypto_ahash_digestsize(tfm); | 
 |  | 
 | 	switch (rctx->dig_sz) { | 
 | #ifdef CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 | 
 | 	case SHA224_DIGEST_SIZE: | 
 | 		rctx->blk_sz = SHA224_BLOCK_SIZE; | 
 | 		rctx->algo = OCS_HCU_ALGO_SHA224; | 
 | 		break; | 
 | #endif /* CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 */ | 
 | 	case SHA256_DIGEST_SIZE: | 
 | 		rctx->blk_sz = SHA256_BLOCK_SIZE; | 
 | 		/* | 
 | 		 * SHA256 and SM3 have the same digest size: use info from tfm | 
 | 		 * context to find out which one we should use. | 
 | 		 */ | 
 | 		rctx->algo = ctx->is_sm3_tfm ? OCS_HCU_ALGO_SM3 : | 
 | 					       OCS_HCU_ALGO_SHA256; | 
 | 		break; | 
 | 	case SHA384_DIGEST_SIZE: | 
 | 		rctx->blk_sz = SHA384_BLOCK_SIZE; | 
 | 		rctx->algo = OCS_HCU_ALGO_SHA384; | 
 | 		break; | 
 | 	case SHA512_DIGEST_SIZE: | 
 | 		rctx->blk_sz = SHA512_BLOCK_SIZE; | 
 | 		rctx->algo = OCS_HCU_ALGO_SHA512; | 
 | 		break; | 
 | 	default: | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	/* Initialize intermediate data. */ | 
 | 	ocs_hcu_hash_init(&rctx->hash_ctx, rctx->algo); | 
 |  | 
 | 	/* If this a HMAC request, set HMAC flag. */ | 
 | 	if (ctx->is_hmac_tfm) | 
 | 		rctx->flags |= REQ_FLAGS_HMAC; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_update(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 | 	int rc; | 
 |  | 
 | 	if (!req->nbytes) | 
 | 		return 0; | 
 |  | 
 | 	rctx->sg_data_total = req->nbytes; | 
 | 	rctx->sg_data_offset = 0; | 
 | 	rctx->sg = req->src; | 
 |  | 
 | 	/* | 
 | 	 * If we are doing HMAC, then we must use SW-assisted HMAC, since HW | 
 | 	 * HMAC does not support context switching (there it can only be used | 
 | 	 * with finup() or digest()). | 
 | 	 */ | 
 | 	if (rctx->flags & REQ_FLAGS_HMAC && | 
 | 	    !(rctx->flags & REQ_FLAGS_HMAC_SW)) { | 
 | 		rctx->flags |= REQ_FLAGS_HMAC_SW; | 
 | 		rc = prepare_ipad(req); | 
 | 		if (rc) | 
 | 			return rc; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * If remaining sg_data fits into ctx buffer, just copy it there; we'll | 
 | 	 * process it at the next update() or final(). | 
 | 	 */ | 
 | 	if (rctx->sg_data_total <= (sizeof(rctx->buffer) - rctx->buf_cnt)) | 
 | 		return flush_sg_to_ocs_buffer(rctx); | 
 |  | 
 | 	return kmb_ocs_hcu_handle_queue(req); | 
 | } | 
 |  | 
 | /* Common logic for kmb_ocs_hcu_final() and kmb_ocs_hcu_finup(). */ | 
 | static int kmb_ocs_hcu_fin_common(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 | 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); | 
 | 	struct ocs_hcu_ctx *ctx = crypto_ahash_ctx(tfm); | 
 | 	int rc; | 
 |  | 
 | 	rctx->flags |= REQ_FINAL; | 
 |  | 
 | 	/* | 
 | 	 * If this is a HMAC request and, so far, we didn't have to switch to | 
 | 	 * SW HMAC, check if we can use HW HMAC. | 
 | 	 */ | 
 | 	if (rctx->flags & REQ_FLAGS_HMAC && | 
 | 	    !(rctx->flags & REQ_FLAGS_HMAC_SW)) { | 
 | 		/* | 
 | 		 * If we are here, it means we never processed any data so far, | 
 | 		 * so we can use HW HMAC, but only if there is some data to | 
 | 		 * process (since OCS HW MAC does not support zero-length | 
 | 		 * messages) and the key length is supported by the hardware | 
 | 		 * (OCS HCU HW only supports length <= 64); if HW HMAC cannot | 
 | 		 * be used, fall back to SW-assisted HMAC. | 
 | 		 */ | 
 | 		if (kmb_get_total_data(rctx) && | 
 | 		    ctx->key_len <= OCS_HCU_HW_KEY_LEN) { | 
 | 			rctx->flags |= REQ_FLAGS_HMAC_HW; | 
 | 		} else { | 
 | 			rctx->flags |= REQ_FLAGS_HMAC_SW; | 
 | 			rc = prepare_ipad(req); | 
 | 			if (rc) | 
 | 				return rc; | 
 | 		} | 
 | 	} | 
 |  | 
 | 	return kmb_ocs_hcu_handle_queue(req); | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_final(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 |  | 
 | 	rctx->sg_data_total = 0; | 
 | 	rctx->sg_data_offset = 0; | 
 | 	rctx->sg = NULL; | 
 |  | 
 | 	return kmb_ocs_hcu_fin_common(req); | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_finup(struct ahash_request *req) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 |  | 
 | 	rctx->sg_data_total = req->nbytes; | 
 | 	rctx->sg_data_offset = 0; | 
 | 	rctx->sg = req->src; | 
 |  | 
 | 	return kmb_ocs_hcu_fin_common(req); | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_digest(struct ahash_request *req) | 
 | { | 
 | 	int rc = 0; | 
 | 	struct ocs_hcu_dev *hcu_dev = kmb_ocs_hcu_find_dev(req); | 
 |  | 
 | 	if (!hcu_dev) | 
 | 		return -ENOENT; | 
 |  | 
 | 	rc = kmb_ocs_hcu_init(req); | 
 | 	if (rc) | 
 | 		return rc; | 
 |  | 
 | 	rc = kmb_ocs_hcu_finup(req); | 
 |  | 
 | 	return rc; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_export(struct ahash_request *req, void *out) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 |  | 
 | 	/* Intermediate data is always stored and applied per request. */ | 
 | 	memcpy(out, rctx, sizeof(*rctx)); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_import(struct ahash_request *req, const void *in) | 
 | { | 
 | 	struct ocs_hcu_rctx *rctx = ahash_request_ctx_dma(req); | 
 |  | 
 | 	/* Intermediate data is always stored and applied per request. */ | 
 | 	memcpy(rctx, in, sizeof(*rctx)); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_setkey(struct crypto_ahash *tfm, const u8 *key, | 
 | 			      unsigned int keylen) | 
 | { | 
 | 	unsigned int digestsize = crypto_ahash_digestsize(tfm); | 
 | 	struct ocs_hcu_ctx *ctx = crypto_ahash_ctx(tfm); | 
 | 	size_t blk_sz = crypto_ahash_blocksize(tfm); | 
 | 	struct crypto_ahash *ahash_tfm; | 
 | 	struct ahash_request *req; | 
 | 	struct crypto_wait wait; | 
 | 	struct scatterlist sg; | 
 | 	const char *alg_name; | 
 | 	int rc; | 
 |  | 
 | 	/* | 
 | 	 * Key length must be equal to block size: | 
 | 	 * - If key is shorter, we are done for now (the key will be padded | 
 | 	 *   later on); this is to maximize the use of HW HMAC (which works | 
 | 	 *   only for keys <= 64 bytes). | 
 | 	 * - If key is longer, we hash it. | 
 | 	 */ | 
 | 	if (keylen <= blk_sz) { | 
 | 		memcpy(ctx->key, key, keylen); | 
 | 		ctx->key_len = keylen; | 
 | 		return 0; | 
 | 	} | 
 |  | 
 | 	switch (digestsize) { | 
 | #ifdef CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 | 
 | 	case SHA224_DIGEST_SIZE: | 
 | 		alg_name = "sha224-keembay-ocs"; | 
 | 		break; | 
 | #endif /* CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 */ | 
 | 	case SHA256_DIGEST_SIZE: | 
 | 		alg_name = ctx->is_sm3_tfm ? "sm3-keembay-ocs" : | 
 | 					     "sha256-keembay-ocs"; | 
 | 		break; | 
 | 	case SHA384_DIGEST_SIZE: | 
 | 		alg_name = "sha384-keembay-ocs"; | 
 | 		break; | 
 | 	case SHA512_DIGEST_SIZE: | 
 | 		alg_name = "sha512-keembay-ocs"; | 
 | 		break; | 
 | 	default: | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	ahash_tfm = crypto_alloc_ahash(alg_name, 0, 0); | 
 | 	if (IS_ERR(ahash_tfm)) | 
 | 		return PTR_ERR(ahash_tfm); | 
 |  | 
 | 	req = ahash_request_alloc(ahash_tfm, GFP_KERNEL); | 
 | 	if (!req) { | 
 | 		rc = -ENOMEM; | 
 | 		goto err_free_ahash; | 
 | 	} | 
 |  | 
 | 	crypto_init_wait(&wait); | 
 | 	ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG, | 
 | 				   crypto_req_done, &wait); | 
 | 	crypto_ahash_clear_flags(ahash_tfm, ~0); | 
 |  | 
 | 	sg_init_one(&sg, key, keylen); | 
 | 	ahash_request_set_crypt(req, &sg, ctx->key, keylen); | 
 |  | 
 | 	rc = crypto_wait_req(crypto_ahash_digest(req), &wait); | 
 | 	if (rc == 0) | 
 | 		ctx->key_len = digestsize; | 
 |  | 
 | 	ahash_request_free(req); | 
 | err_free_ahash: | 
 | 	crypto_free_ahash(ahash_tfm); | 
 |  | 
 | 	return rc; | 
 | } | 
 |  | 
 | /* Set request size and initialize tfm context. */ | 
 | static void __cra_init(struct crypto_tfm *tfm, struct ocs_hcu_ctx *ctx) | 
 | { | 
 | 	crypto_ahash_set_reqsize_dma(__crypto_ahash_cast(tfm), | 
 | 				     sizeof(struct ocs_hcu_rctx)); | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_sha_cra_init(struct crypto_tfm *tfm) | 
 | { | 
 | 	struct ocs_hcu_ctx *ctx = crypto_tfm_ctx(tfm); | 
 |  | 
 | 	__cra_init(tfm, ctx); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_sm3_cra_init(struct crypto_tfm *tfm) | 
 | { | 
 | 	struct ocs_hcu_ctx *ctx = crypto_tfm_ctx(tfm); | 
 |  | 
 | 	__cra_init(tfm, ctx); | 
 |  | 
 | 	ctx->is_sm3_tfm = true; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_hmac_sm3_cra_init(struct crypto_tfm *tfm) | 
 | { | 
 | 	struct ocs_hcu_ctx *ctx = crypto_tfm_ctx(tfm); | 
 |  | 
 | 	__cra_init(tfm, ctx); | 
 |  | 
 | 	ctx->is_sm3_tfm = true; | 
 | 	ctx->is_hmac_tfm = true; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_hmac_cra_init(struct crypto_tfm *tfm) | 
 | { | 
 | 	struct ocs_hcu_ctx *ctx = crypto_tfm_ctx(tfm); | 
 |  | 
 | 	__cra_init(tfm, ctx); | 
 |  | 
 | 	ctx->is_hmac_tfm = true; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | /* Function called when 'tfm' is de-initialized. */ | 
 | static void kmb_ocs_hcu_hmac_cra_exit(struct crypto_tfm *tfm) | 
 | { | 
 | 	struct ocs_hcu_ctx *ctx = crypto_tfm_ctx(tfm); | 
 |  | 
 | 	/* Clear the key. */ | 
 | 	memzero_explicit(ctx->key, sizeof(ctx->key)); | 
 | } | 
 |  | 
 | static struct ahash_engine_alg ocs_hcu_algs[] = { | 
 | #ifdef CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA224_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "sha224", | 
 | 			.cra_driver_name	= "sha224-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA224_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_sha_cra_init, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.setkey		= kmb_ocs_hcu_setkey, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA224_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "hmac(sha224)", | 
 | 			.cra_driver_name	= "hmac-sha224-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA224_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_hmac_cra_init, | 
 | 			.cra_exit		= kmb_ocs_hcu_hmac_cra_exit, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | #endif /* CONFIG_CRYPTO_DEV_KEEMBAY_OCS_HCU_HMAC_SHA224 */ | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA256_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "sha256", | 
 | 			.cra_driver_name	= "sha256-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA256_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_sha_cra_init, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.setkey		= kmb_ocs_hcu_setkey, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA256_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "hmac(sha256)", | 
 | 			.cra_driver_name	= "hmac-sha256-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA256_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_hmac_cra_init, | 
 | 			.cra_exit		= kmb_ocs_hcu_hmac_cra_exit, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SM3_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "sm3", | 
 | 			.cra_driver_name	= "sm3-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SM3_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_sm3_cra_init, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.setkey		= kmb_ocs_hcu_setkey, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SM3_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "hmac(sm3)", | 
 | 			.cra_driver_name	= "hmac-sm3-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SM3_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_hmac_sm3_cra_init, | 
 | 			.cra_exit		= kmb_ocs_hcu_hmac_cra_exit, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA384_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "sha384", | 
 | 			.cra_driver_name	= "sha384-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA384_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_sha_cra_init, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.setkey		= kmb_ocs_hcu_setkey, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA384_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "hmac(sha384)", | 
 | 			.cra_driver_name	= "hmac-sha384-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA384_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_hmac_cra_init, | 
 | 			.cra_exit		= kmb_ocs_hcu_hmac_cra_exit, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA512_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "sha512", | 
 | 			.cra_driver_name	= "sha512-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA512_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_sha_cra_init, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | { | 
 | 	.base.init		= kmb_ocs_hcu_init, | 
 | 	.base.update		= kmb_ocs_hcu_update, | 
 | 	.base.final		= kmb_ocs_hcu_final, | 
 | 	.base.finup		= kmb_ocs_hcu_finup, | 
 | 	.base.digest		= kmb_ocs_hcu_digest, | 
 | 	.base.export		= kmb_ocs_hcu_export, | 
 | 	.base.import		= kmb_ocs_hcu_import, | 
 | 	.base.setkey		= kmb_ocs_hcu_setkey, | 
 | 	.base.halg = { | 
 | 		.digestsize	= SHA512_DIGEST_SIZE, | 
 | 		.statesize	= sizeof(struct ocs_hcu_rctx), | 
 | 		.base	= { | 
 | 			.cra_name		= "hmac(sha512)", | 
 | 			.cra_driver_name	= "hmac-sha512-keembay-ocs", | 
 | 			.cra_priority		= 255, | 
 | 			.cra_flags		= CRYPTO_ALG_ASYNC, | 
 | 			.cra_blocksize		= SHA512_BLOCK_SIZE, | 
 | 			.cra_ctxsize		= sizeof(struct ocs_hcu_ctx), | 
 | 			.cra_alignmask		= 0, | 
 | 			.cra_module		= THIS_MODULE, | 
 | 			.cra_init		= kmb_ocs_hcu_hmac_cra_init, | 
 | 			.cra_exit		= kmb_ocs_hcu_hmac_cra_exit, | 
 | 		} | 
 | 	}, | 
 | 	.op.do_one_request = kmb_ocs_hcu_do_one_request, | 
 | }, | 
 | }; | 
 |  | 
 | /* Device tree driver match. */ | 
 | static const struct of_device_id kmb_ocs_hcu_of_match[] = { | 
 | 	{ | 
 | 		.compatible = "intel,keembay-ocs-hcu", | 
 | 	}, | 
 | 	{} | 
 | }; | 
 | MODULE_DEVICE_TABLE(of, kmb_ocs_hcu_of_match); | 
 |  | 
 | static void kmb_ocs_hcu_remove(struct platform_device *pdev) | 
 | { | 
 | 	struct ocs_hcu_dev *hcu_dev = platform_get_drvdata(pdev); | 
 |  | 
 | 	crypto_engine_unregister_ahashes(ocs_hcu_algs, ARRAY_SIZE(ocs_hcu_algs)); | 
 |  | 
 | 	crypto_engine_exit(hcu_dev->engine); | 
 |  | 
 | 	spin_lock_bh(&ocs_hcu.lock); | 
 | 	list_del(&hcu_dev->list); | 
 | 	spin_unlock_bh(&ocs_hcu.lock); | 
 | } | 
 |  | 
 | static int kmb_ocs_hcu_probe(struct platform_device *pdev) | 
 | { | 
 | 	struct device *dev = &pdev->dev; | 
 | 	struct ocs_hcu_dev *hcu_dev; | 
 | 	int rc; | 
 |  | 
 | 	hcu_dev = devm_kzalloc(dev, sizeof(*hcu_dev), GFP_KERNEL); | 
 | 	if (!hcu_dev) | 
 | 		return -ENOMEM; | 
 |  | 
 | 	hcu_dev->dev = dev; | 
 |  | 
 | 	platform_set_drvdata(pdev, hcu_dev); | 
 | 	rc = dma_set_mask_and_coherent(&pdev->dev, OCS_HCU_DMA_BIT_MASK); | 
 | 	if (rc) | 
 | 		return rc; | 
 |  | 
 | 	hcu_dev->io_base = devm_platform_ioremap_resource(pdev, 0); | 
 | 	if (IS_ERR(hcu_dev->io_base)) | 
 | 		return PTR_ERR(hcu_dev->io_base); | 
 |  | 
 | 	init_completion(&hcu_dev->irq_done); | 
 |  | 
 | 	/* Get and request IRQ. */ | 
 | 	hcu_dev->irq = platform_get_irq(pdev, 0); | 
 | 	if (hcu_dev->irq < 0) | 
 | 		return hcu_dev->irq; | 
 |  | 
 | 	rc = devm_request_threaded_irq(&pdev->dev, hcu_dev->irq, | 
 | 				       ocs_hcu_irq_handler, NULL, 0, | 
 | 				       "keembay-ocs-hcu", hcu_dev); | 
 | 	if (rc < 0) { | 
 | 		dev_err(dev, "Could not request IRQ.\n"); | 
 | 		return rc; | 
 | 	} | 
 |  | 
 | 	INIT_LIST_HEAD(&hcu_dev->list); | 
 |  | 
 | 	spin_lock_bh(&ocs_hcu.lock); | 
 | 	list_add_tail(&hcu_dev->list, &ocs_hcu.dev_list); | 
 | 	spin_unlock_bh(&ocs_hcu.lock); | 
 |  | 
 | 	/* Initialize crypto engine */ | 
 | 	hcu_dev->engine = crypto_engine_alloc_init(dev, 1); | 
 | 	if (!hcu_dev->engine) { | 
 | 		rc = -ENOMEM; | 
 | 		goto list_del; | 
 | 	} | 
 |  | 
 | 	rc = crypto_engine_start(hcu_dev->engine); | 
 | 	if (rc) { | 
 | 		dev_err(dev, "Could not start engine.\n"); | 
 | 		goto cleanup; | 
 | 	} | 
 |  | 
 | 	/* Security infrastructure guarantees OCS clock is enabled. */ | 
 |  | 
 | 	rc = crypto_engine_register_ahashes(ocs_hcu_algs, ARRAY_SIZE(ocs_hcu_algs)); | 
 | 	if (rc) { | 
 | 		dev_err(dev, "Could not register algorithms.\n"); | 
 | 		goto cleanup; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 |  | 
 | cleanup: | 
 | 	crypto_engine_exit(hcu_dev->engine); | 
 | list_del: | 
 | 	spin_lock_bh(&ocs_hcu.lock); | 
 | 	list_del(&hcu_dev->list); | 
 | 	spin_unlock_bh(&ocs_hcu.lock); | 
 |  | 
 | 	return rc; | 
 | } | 
 |  | 
 | /* The OCS driver is a platform device. */ | 
 | static struct platform_driver kmb_ocs_hcu_driver = { | 
 | 	.probe = kmb_ocs_hcu_probe, | 
 | 	.remove_new = kmb_ocs_hcu_remove, | 
 | 	.driver = { | 
 | 			.name = DRV_NAME, | 
 | 			.of_match_table = kmb_ocs_hcu_of_match, | 
 | 		}, | 
 | }; | 
 |  | 
 | module_platform_driver(kmb_ocs_hcu_driver); | 
 |  | 
 | MODULE_LICENSE("GPL"); |