mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-10-30 16:07:39 +00:00
crypto: ccree - remove unneeded casts
Unneeded casts prevent the compiler from performing valuable checks. This is especially true for function pointers. Remove these casts, to prevent silently introducing bugs when a variable's type might be changed in the future. No change in generated code. Signed-off-by: Geert Uytterhoeven <geert+renesas@glider.be> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
b83fd3e5ec
commit
f4274eeca4
4 changed files with 23 additions and 25 deletions
|
@ -448,8 +448,7 @@ static int cc_get_plain_hmac_key(struct crypto_aead *tfm, const u8 *authkey,
|
|||
if (!key)
|
||||
return -ENOMEM;
|
||||
|
||||
key_dma_addr = dma_map_single(dev, (void *)key, keylen,
|
||||
DMA_TO_DEVICE);
|
||||
key_dma_addr = dma_map_single(dev, key, keylen, DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(dev, key_dma_addr)) {
|
||||
dev_err(dev, "Mapping key va=0x%p len=%u for DMA failed\n",
|
||||
key, keylen);
|
||||
|
@ -1921,8 +1920,8 @@ static int cc_proc_aead(struct aead_request *req,
|
|||
}
|
||||
|
||||
/* Setup request structure */
|
||||
cc_req.user_cb = (void *)cc_aead_complete;
|
||||
cc_req.user_arg = (void *)req;
|
||||
cc_req.user_cb = cc_aead_complete;
|
||||
cc_req.user_arg = req;
|
||||
|
||||
/* Setup request context */
|
||||
areq_ctx->gen_ctx.op_type = direct;
|
||||
|
|
|
@ -118,7 +118,7 @@ void cc_copy_sg_portion(struct device *dev, u8 *dest, struct scatterlist *sg,
|
|||
u32 nents;
|
||||
|
||||
nents = sg_nents_for_len(sg, end);
|
||||
sg_copy_buffer(sg, nents, (void *)dest, (end - to_skip + 1), to_skip,
|
||||
sg_copy_buffer(sg, nents, dest, (end - to_skip + 1), to_skip,
|
||||
(direct == CC_SG_TO_BUF));
|
||||
}
|
||||
|
||||
|
@ -415,10 +415,9 @@ int cc_map_cipher_request(struct cc_drvdata *drvdata, void *ctx,
|
|||
|
||||
/* Map IV buffer */
|
||||
if (ivsize) {
|
||||
dump_byte_array("iv", (u8 *)info, ivsize);
|
||||
dump_byte_array("iv", info, ivsize);
|
||||
req_ctx->gen_ctx.iv_dma_addr =
|
||||
dma_map_single(dev, (void *)info,
|
||||
ivsize, DMA_BIDIRECTIONAL);
|
||||
dma_map_single(dev, info, ivsize, DMA_BIDIRECTIONAL);
|
||||
if (dma_mapping_error(dev, req_ctx->gen_ctx.iv_dma_addr)) {
|
||||
dev_err(dev, "Mapping iv %u B at va=%pK for DMA failed\n",
|
||||
ivsize, info);
|
||||
|
|
|
@ -184,7 +184,7 @@ static int cc_cipher_init(struct crypto_tfm *tfm)
|
|||
ctx_p->user.key);
|
||||
|
||||
/* Map key buffer */
|
||||
ctx_p->user.key_dma_addr = dma_map_single(dev, (void *)ctx_p->user.key,
|
||||
ctx_p->user.key_dma_addr = dma_map_single(dev, ctx_p->user.key,
|
||||
max_key_buf_size,
|
||||
DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(dev, ctx_p->user.key_dma_addr)) {
|
||||
|
@ -284,7 +284,7 @@ static int cc_cipher_sethkey(struct crypto_skcipher *sktfm, const u8 *key,
|
|||
|
||||
dev_dbg(dev, "Setting HW key in context @%p for %s. keylen=%u\n",
|
||||
ctx_p, crypto_tfm_alg_name(tfm), keylen);
|
||||
dump_byte_array("key", (u8 *)key, keylen);
|
||||
dump_byte_array("key", key, keylen);
|
||||
|
||||
/* STAT_PHASE_0: Init and sanity checks */
|
||||
|
||||
|
@ -387,7 +387,7 @@ static int cc_cipher_setkey(struct crypto_skcipher *sktfm, const u8 *key,
|
|||
|
||||
dev_dbg(dev, "Setting key in context @%p for %s. keylen=%u\n",
|
||||
ctx_p, crypto_tfm_alg_name(tfm), keylen);
|
||||
dump_byte_array("key", (u8 *)key, keylen);
|
||||
dump_byte_array("key", key, keylen);
|
||||
|
||||
/* STAT_PHASE_0: Init and sanity checks */
|
||||
|
||||
|
@ -885,8 +885,8 @@ static int cc_cipher_process(struct skcipher_request *req,
|
|||
}
|
||||
|
||||
/* Setup request structure */
|
||||
cc_req.user_cb = (void *)cc_cipher_complete;
|
||||
cc_req.user_arg = (void *)req;
|
||||
cc_req.user_cb = cc_cipher_complete;
|
||||
cc_req.user_arg = req;
|
||||
|
||||
/* Setup CPP operation details */
|
||||
if (ctx_p->key_type == CC_POLICY_PROTECTED_KEY) {
|
||||
|
|
|
@ -752,7 +752,7 @@ static int cc_hash_setkey(struct crypto_ahash *ahash, const u8 *key,
|
|||
return -ENOMEM;
|
||||
|
||||
ctx->key_params.key_dma_addr =
|
||||
dma_map_single(dev, (void *)ctx->key_params.key, keylen,
|
||||
dma_map_single(dev, ctx->key_params.key, keylen,
|
||||
DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(dev, ctx->key_params.key_dma_addr)) {
|
||||
dev_err(dev, "Mapping key va=0x%p len=%u for DMA failed\n",
|
||||
|
@ -1067,8 +1067,8 @@ static int cc_alloc_ctx(struct cc_hash_ctx *ctx)
|
|||
ctx->key_params.keylen = 0;
|
||||
|
||||
ctx->digest_buff_dma_addr =
|
||||
dma_map_single(dev, (void *)ctx->digest_buff,
|
||||
sizeof(ctx->digest_buff), DMA_BIDIRECTIONAL);
|
||||
dma_map_single(dev, ctx->digest_buff, sizeof(ctx->digest_buff),
|
||||
DMA_BIDIRECTIONAL);
|
||||
if (dma_mapping_error(dev, ctx->digest_buff_dma_addr)) {
|
||||
dev_err(dev, "Mapping digest len %zu B at va=%pK for DMA failed\n",
|
||||
sizeof(ctx->digest_buff), ctx->digest_buff);
|
||||
|
@ -1079,7 +1079,7 @@ static int cc_alloc_ctx(struct cc_hash_ctx *ctx)
|
|||
&ctx->digest_buff_dma_addr);
|
||||
|
||||
ctx->opad_tmp_keys_dma_addr =
|
||||
dma_map_single(dev, (void *)ctx->opad_tmp_keys_buff,
|
||||
dma_map_single(dev, ctx->opad_tmp_keys_buff,
|
||||
sizeof(ctx->opad_tmp_keys_buff),
|
||||
DMA_BIDIRECTIONAL);
|
||||
if (dma_mapping_error(dev, ctx->opad_tmp_keys_dma_addr)) {
|
||||
|
@ -1196,8 +1196,8 @@ static int cc_mac_update(struct ahash_request *req)
|
|||
idx++;
|
||||
|
||||
/* Setup request structure */
|
||||
cc_req.user_cb = (void *)cc_update_complete;
|
||||
cc_req.user_arg = (void *)req;
|
||||
cc_req.user_cb = cc_update_complete;
|
||||
cc_req.user_arg = req;
|
||||
|
||||
rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base);
|
||||
if (rc != -EINPROGRESS && rc != -EBUSY) {
|
||||
|
@ -1254,8 +1254,8 @@ static int cc_mac_final(struct ahash_request *req)
|
|||
}
|
||||
|
||||
/* Setup request structure */
|
||||
cc_req.user_cb = (void *)cc_hash_complete;
|
||||
cc_req.user_arg = (void *)req;
|
||||
cc_req.user_cb = cc_hash_complete;
|
||||
cc_req.user_arg = req;
|
||||
|
||||
if (state->xcbc_count && rem_cnt == 0) {
|
||||
/* Load key for ECB decryption */
|
||||
|
@ -1369,8 +1369,8 @@ static int cc_mac_finup(struct ahash_request *req)
|
|||
}
|
||||
|
||||
/* Setup request structure */
|
||||
cc_req.user_cb = (void *)cc_hash_complete;
|
||||
cc_req.user_arg = (void *)req;
|
||||
cc_req.user_cb = cc_hash_complete;
|
||||
cc_req.user_arg = req;
|
||||
|
||||
if (ctx->hw_mode == DRV_CIPHER_XCBC_MAC) {
|
||||
key_len = CC_AES_128_BIT_KEY_SIZE;
|
||||
|
@ -1448,8 +1448,8 @@ static int cc_mac_digest(struct ahash_request *req)
|
|||
}
|
||||
|
||||
/* Setup request structure */
|
||||
cc_req.user_cb = (void *)cc_digest_complete;
|
||||
cc_req.user_arg = (void *)req;
|
||||
cc_req.user_cb = cc_digest_complete;
|
||||
cc_req.user_arg = req;
|
||||
|
||||
if (ctx->hw_mode == DRV_CIPHER_XCBC_MAC) {
|
||||
key_len = CC_AES_128_BIT_KEY_SIZE;
|
||||
|
|
Loading…
Reference in a new issue