diff options
-rw-r--r-- | drivers/cpufreq/qoriq-cpufreq.c | 4 | ||||
-rw-r--r-- | drivers/crypto/caam/caamalg.c | 12 | ||||
-rw-r--r-- | drivers/crypto/caam/caamhash.c | 43 | ||||
-rw-r--r-- | drivers/staging/fsl_dce/tests/chunk_comp_def_cf/dce_chunk_comp_def_cf_test.c | 171 | ||||
-rw-r--r-- | drivers/staging/fsl_dce/tests/performance_simple/dce_perf_simple.c | 8 | ||||
-rw-r--r-- | drivers/staging/fsl_dce/tests/performance_simple/dce_sf_perf_simple.c | 13 | ||||
-rw-r--r-- | drivers/staging/fsl_pme2/pme2_ctrl.c | 4 | ||||
-rw-r--r-- | drivers/staging/fsl_pme2/pme2_db.c | 1 | ||||
-rw-r--r-- | drivers/staging/fsl_pme2/pme2_high.c | 2 | ||||
-rw-r--r-- | drivers/staging/fsl_qbman/qman_high.c | 12 | ||||
-rw-r--r-- | net/core/dev.c | 9 |
11 files changed, 181 insertions, 98 deletions
diff --git a/drivers/cpufreq/qoriq-cpufreq.c b/drivers/cpufreq/qoriq-cpufreq.c index 36b2e6e..cd620fb 100644 --- a/drivers/cpufreq/qoriq-cpufreq.c +++ b/drivers/cpufreq/qoriq-cpufreq.c @@ -525,7 +525,7 @@ static int __init qoriq_cpufreq_init(void) ret = cpufreq_register_driver(&qoriq_cpufreq_driver); if (!ret) - pr_info("Freescale PowerPC qoriq CPU frequency scaling driver\n"); + pr_info("Freescale QorIQ CPU frequency scaling driver\n"); return ret; } @@ -539,4 +539,4 @@ module_exit(qoriq_cpufreq_exit); MODULE_LICENSE("GPL"); MODULE_AUTHOR("Tang Yuantian <Yuantian.Tang@freescale.com>"); -MODULE_DESCRIPTION("cpufreq driver for Freescale e500mc series SoCs"); +MODULE_DESCRIPTION("cpufreq driver for Freescale QorIQ series SoCs"); diff --git a/drivers/crypto/caam/caamalg.c b/drivers/crypto/caam/caamalg.c index 0f117d0..023a5d8 100644 --- a/drivers/crypto/caam/caamalg.c +++ b/drivers/crypto/caam/caamalg.c @@ -3070,8 +3070,8 @@ static struct aead_edesc *aead_edesc_alloc(struct aead_request *req, sec4_sg_bytes = sec4_sg_len * sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct aead_edesc) + desc_bytes + - sec4_sg_bytes, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return ERR_PTR(-ENOMEM); @@ -3310,8 +3310,8 @@ static struct aead_edesc *aead_giv_edesc_alloc(struct aead_givcrypt_request sec4_sg_bytes = sec4_sg_len * sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct aead_edesc) + desc_bytes + - sec4_sg_bytes, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return ERR_PTR(-ENOMEM); @@ -3581,8 +3581,8 @@ static struct ablkcipher_edesc *ablkcipher_edesc_alloc(struct ablkcipher_request sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct ablkcipher_edesc) + desc_bytes + - sec4_sg_bytes, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + desc_bytes + sec4_sg_bytes, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return ERR_PTR(-ENOMEM); diff --git a/drivers/crypto/caam/caamhash.c b/drivers/crypto/caam/caamhash.c index e99a45b..fb913c3 100644 --- a/drivers/crypto/caam/caamhash.c +++ b/drivers/crypto/caam/caamhash.c @@ -807,7 +807,7 @@ static int ahash_update_ctx(struct ahash_request *req) * allocate space for base edesc and hw desc commands, * link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN + + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes, GFP_DMA | flags); if (!edesc) { dev_err(jrdev, @@ -835,17 +835,17 @@ static int ahash_update_ctx(struct ahash_request *req) src_map_to_sec4_sg(jrdev, req->src, src_nents, edesc->sec4_sg + sec4_sg_src_index, chained); - if (*next_buflen) { + if (*next_buflen) scatterwalk_map_and_copy(next_buf, req->src, to_hash - *buflen, *next_buflen, 0); - state->current_buf = !state->current_buf; - } } else { (edesc->sec4_sg + sec4_sg_src_index - 1)->len |= SEC4_SG_LEN_FIN; } + state->current_buf = !state->current_buf; + sh_len = desc_len(sh_desc); desc = edesc->hw_desc; init_job_desc_shared(desc, ptr, sh_len, HDR_SHARE_DEFER | @@ -909,17 +909,18 @@ static int ahash_final_ctx(struct ahash_request *req) state->buflen_1; u32 *sh_desc = ctx->sh_desc_fin, *desc; dma_addr_t ptr = ctx->sh_desc_fin_dma; - int sec4_sg_bytes; + int sec4_sg_bytes, sec4_sg_src_index; int digestsize = crypto_ahash_digestsize(ahash); struct ahash_edesc *edesc; int ret = 0; int sh_len; - sec4_sg_bytes = (1 + (buflen ? 1 : 0)) * sizeof(struct sec4_sg_entry); + sec4_sg_src_index = 1 + (buflen ? 1 : 0); + sec4_sg_bytes = sec4_sg_src_index * sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN + - sec4_sg_bytes, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return -ENOMEM; @@ -942,7 +943,7 @@ static int ahash_final_ctx(struct ahash_request *req) state->buf_dma = try_buf_map_to_sec4_sg(jrdev, edesc->sec4_sg + 1, buf, state->buf_dma, buflen, last_buflen); - (edesc->sec4_sg + sec4_sg_bytes - 1)->len |= SEC4_SG_LEN_FIN; + (edesc->sec4_sg + sec4_sg_src_index - 1)->len |= SEC4_SG_LEN_FIN; edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg, sec4_sg_bytes, DMA_TO_DEVICE); @@ -1005,8 +1006,8 @@ static int ahash_finup_ctx(struct ahash_request *req) sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN + - sec4_sg_bytes, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return -ENOMEM; @@ -1091,8 +1092,8 @@ static int ahash_digest(struct ahash_request *req) sec4_sg_bytes = src_nents * sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + sec4_sg_bytes + - DESC_JOB_IO_LEN, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + sec4_sg_bytes + DESC_JOB_IO_LEN, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return -ENOMEM; @@ -1165,8 +1166,7 @@ static int ahash_final_no_ctx(struct ahash_request *req) int sh_len; /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN, - GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN, GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return -ENOMEM; @@ -1245,7 +1245,7 @@ static int ahash_update_no_ctx(struct ahash_request *req) * allocate space for base edesc and hw desc commands, * link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN + + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes, GFP_DMA | flags); if (!edesc) { dev_err(jrdev, @@ -1268,9 +1268,10 @@ static int ahash_update_no_ctx(struct ahash_request *req) scatterwalk_map_and_copy(next_buf, req->src, to_hash - *buflen, *next_buflen, 0); - state->current_buf = !state->current_buf; } + state->current_buf = !state->current_buf; + sh_len = desc_len(sh_desc); desc = edesc->hw_desc; init_job_desc_shared(desc, ptr, sh_len, HDR_SHARE_DEFER | @@ -1352,8 +1353,8 @@ static int ahash_finup_no_ctx(struct ahash_request *req) sizeof(struct sec4_sg_entry); /* allocate space for base edesc and hw desc commands, link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN + - sec4_sg_bytes, GFP_DMA | flags); + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes, + GFP_DMA | flags); if (!edesc) { dev_err(jrdev, "could not allocate extended descriptor\n"); return -ENOMEM; @@ -1447,7 +1448,7 @@ static int ahash_update_first(struct ahash_request *req) * allocate space for base edesc and hw desc commands, * link tables */ - edesc = kmalloc(sizeof(struct ahash_edesc) + DESC_JOB_IO_LEN + + edesc = kzalloc(sizeof(*edesc) + DESC_JOB_IO_LEN + sec4_sg_bytes, GFP_DMA | flags); if (!edesc) { dev_err(jrdev, @@ -1544,6 +1545,8 @@ static int ahash_init(struct ahash_request *req) state->current_buf = 0; state->buf_dma = 0; + state->buflen_0 = 0; + state->buflen_1 = 0; return 0; } diff --git a/drivers/staging/fsl_dce/tests/chunk_comp_def_cf/dce_chunk_comp_def_cf_test.c b/drivers/staging/fsl_dce/tests/chunk_comp_def_cf/dce_chunk_comp_def_cf_test.c index 80e2ca1..7650e08 100644 --- a/drivers/staging/fsl_dce/tests/chunk_comp_def_cf/dce_chunk_comp_def_cf_test.c +++ b/drivers/staging/fsl_dce/tests/chunk_comp_def_cf/dce_chunk_comp_def_cf_test.c @@ -179,7 +179,7 @@ static int do_test(struct dce_test_ctx *ctx, nop_req = kmalloc(sizeof(*nop_req), GFP_KERNEL); if (!nop_req) { ret = -ENOMEM; - goto fail_deflate_nop; + goto fail_nop; } init_completion(&nop_req->cb_done); @@ -188,7 +188,7 @@ static int do_test(struct dce_test_ctx *ctx, ret = fsl_dce_chunk_nop(&ctx->deflate_chunk, 0, nop_req); if (ret) { BUG(); - goto fail_deflate_nop; + goto fail_nop; } pr_info("Sent NOP on deflate path\n"); @@ -199,7 +199,7 @@ static int do_test(struct dce_test_ctx *ctx, ret = fsl_dce_chunk_nop(&ctx->inflate_chunk, 0, nop_req); if (ret) { BUG(); - goto fail_inflate_nop; + goto fail_nop; } pr_info("Sent NOP on inflate path\n"); @@ -212,14 +212,16 @@ static int do_test(struct dce_test_ctx *ctx, def_process_req = kzalloc(sizeof(*def_process_req), GFP_KERNEL); if (!def_process_req) { BUG(); - goto fail_inflate_params; + ret = -ENOMEM; + goto fail_nop; } pr_info("Allocated def_process_req\n"); def_process_req->v_output = vmalloc(output_len); - if (!def_process_req) { + if (!def_process_req->v_output) { BUG(); - goto fail_v_output; + ret = -ENOMEM; + goto fail_deflate_v_output; } init_completion(&def_process_req->cb_done); @@ -228,7 +230,7 @@ static int do_test(struct dce_test_ctx *ctx, &def_process_req->input_data); if (ret) { BUG(); - goto fail_alloc_dce_data_input; + goto fail_deflate_alloc_dce_data_input; } if (verbose_level == 1) { @@ -240,7 +242,7 @@ static int do_test(struct dce_test_ctx *ctx, &def_process_req->output_data); if (ret) { BUG(); - goto fail_alloc_dce_data_output; + goto fail_deflate_alloc_dce_data_output; } if (verbose_level == 1) { @@ -252,7 +254,7 @@ static int do_test(struct dce_test_ctx *ctx, &def_process_req->input_data); if (ret) { BUG(); - goto fail_alloc_dce_data_output; + goto fail_deflate_copy_input_to_dce_data; } if (verbose_level == 1) { @@ -263,7 +265,7 @@ static int do_test(struct dce_test_ctx *ctx, ret = dma_map_dce_data(&def_process_req->input_data, DMA_BIDIRECTIONAL); if (ret) { BUG(); - goto fail_alloc_dce_data_output; + goto fail_deflate_dma_map_dce_data_input; } if (verbose_level == 1) { @@ -275,7 +277,7 @@ static int do_test(struct dce_test_ctx *ctx, DMA_BIDIRECTIONAL); if (ret) { BUG(); - goto fail_dma_map_deflate_output_data; + goto fail_deflate_dma_map_dce_data_output; } if (verbose_level == 1) { @@ -288,7 +290,7 @@ static int do_test(struct dce_test_ctx *ctx, DMA_BIDIRECTIONAL); if (ret) { BUG(); - goto fail_output_attach_data_to_sg_deflate; + goto fail_deflate_attach_data_list_to_sg_output; } ret = attach_data_list_to_sg(&def_process_req->dce_cf[1], @@ -296,7 +298,7 @@ static int do_test(struct dce_test_ctx *ctx, DMA_BIDIRECTIONAL); if (ret) { BUG(); - goto fail_input_attach_data_to_sg_deflate; + goto fail_deflate_attach_data_list_to_sg_input; } def_process_req->dce_cf[2].final = 1; @@ -304,7 +306,7 @@ static int do_test(struct dce_test_ctx *ctx, def_process_req->input_fd._format2 = qm_fd_compound; def_process_req->input_fd.cong_weight = 1; qm_fd_addr_set64(&def_process_req->input_fd, - fsl_dce_map(def_process_req->dce_cf)); + fsl_dce_map(def_process_req->dce_cf)); print_dce_fd(def_process_req->input_fd); print_dce_sg(def_process_req->dce_cf[0]); @@ -336,28 +338,28 @@ static int do_test(struct dce_test_ctx *ctx, &def_process_req->input_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - goto fail_input_attach_data_to_sg_deflate; + goto fail_deflate_attach_data_list_to_sg_input; } ret = detach_data_list_from_sg(&def_process_req->dce_cf[0], &def_process_req->output_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - goto fail_output_attach_data_to_sg_deflate; + goto fail_deflate_attach_data_list_to_sg_output; } ret = dma_unmap_dce_data(&def_process_req->output_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - goto fail_dma_map_deflate_output_data; + goto fail_deflate_dma_map_dce_data_output; } ret = dma_unmap_dce_data(&def_process_req->input_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - goto fail_alloc_dce_data_output; + goto fail_deflate_dma_map_dce_data_input; } pr_info("Got chunk process, status = %d, sg_table[0].length = %d\n", @@ -376,8 +378,10 @@ static int do_test(struct dce_test_ctx *ctx, def_process_req->v_output = vmalloc(def_process_req->dce_cf[0].length); if (!def_process_req->v_output) { pr_err("Error %d\n", __LINE__); - goto fail_alloc_dce_data_output; + ret = -ENOMEM; + goto fail_deflate_copy_input_to_dce_data; } + def_process_req->v_output_size = def_process_req->dce_cf[0].length; ret = copy_output_dce_data_to_buffer(&def_process_req->output_data, @@ -386,7 +390,7 @@ static int do_test(struct dce_test_ctx *ctx, if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_deflate_copy_input_to_dce_data; } /* Free dce data deflate operation, but keep vmalloc output */ @@ -398,14 +402,16 @@ static int do_test(struct dce_test_ctx *ctx, inf_process_req = kzalloc(sizeof(*inf_process_req), GFP_KERNEL); if (!inf_process_req) { pr_err("Error %d\n", __LINE__); - return ret; + ret = -ENOMEM; + goto fail_inflate_params; } pr_info("Allocated inf_process_req\n"); inf_process_req->v_output = vmalloc(input_len); - if (!def_process_req) { + if (!inf_process_req->v_output) { pr_err("Error %d\n", __LINE__); - return ret; + ret = -ENOMEM; + goto fail_inflate_v_output; } init_completion(&inf_process_req->cb_done); @@ -415,7 +421,7 @@ static int do_test(struct dce_test_ctx *ctx, &inf_process_req->input_data); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_alloc_dce_data_input; } if (verbose_level == 1) { @@ -427,7 +433,7 @@ static int do_test(struct dce_test_ctx *ctx, &inf_process_req->output_data); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_alloc_dce_data_output; } if (verbose_level == 1) { @@ -439,7 +445,7 @@ static int do_test(struct dce_test_ctx *ctx, def_process_req->v_output_size, &inf_process_req->input_data); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_copy_input_to_dce_data; } if (verbose_level == 1) { @@ -450,7 +456,7 @@ static int do_test(struct dce_test_ctx *ctx, ret = dma_map_dce_data(&inf_process_req->input_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_dma_map_dce_data_input; } if (verbose_level == 1) { @@ -462,7 +468,7 @@ static int do_test(struct dce_test_ctx *ctx, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_dma_map_dce_data_output; } if (verbose_level == 1) { @@ -474,21 +480,21 @@ static int do_test(struct dce_test_ctx *ctx, &inf_process_req->output_data, true, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_attach_data_list_to_sg_output; } ret = attach_data_list_to_sg(&inf_process_req->dce_cf[1], &inf_process_req->input_data, false, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_attach_data_list_to_sg_input; } inf_process_req->dce_cf[2].final = 1; inf_process_req->input_fd._format2 = qm_fd_compound; qm_fd_addr_set64(&inf_process_req->input_fd, - fsl_dce_map(inf_process_req->dce_cf)); + fsl_dce_map(inf_process_req->dce_cf)); print_dce_fd(inf_process_req->input_fd); print_dce_sg(inf_process_req->dce_cf[0]); @@ -503,32 +509,32 @@ static int do_test(struct dce_test_ctx *ctx, pr_info("Output FD\n"); print_dce_fd(inf_process_req->output_fd); - ret = dma_unmap_dce_data(&inf_process_req->input_data, - DMA_BIDIRECTIONAL); + ret = detach_data_list_from_sg(&inf_process_req->dce_cf[1], + &inf_process_req->input_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_attach_data_list_to_sg_input; } - ret = dma_unmap_dce_data(&inf_process_req->output_data, - DMA_BIDIRECTIONAL); + ret = detach_data_list_from_sg(&inf_process_req->dce_cf[0], + &inf_process_req->output_data, DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_attach_data_list_to_sg_output; } - ret = detach_data_list_from_sg(&inf_process_req->dce_cf[0], - &inf_process_req->output_data, DMA_BIDIRECTIONAL); + ret = dma_unmap_dce_data(&inf_process_req->output_data, + DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_dma_map_dce_data_output; } - ret = detach_data_list_from_sg(&inf_process_req->dce_cf[1], - &inf_process_req->input_data, DMA_BIDIRECTIONAL); + ret = dma_unmap_dce_data(&inf_process_req->input_data, + DMA_BIDIRECTIONAL); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_dma_map_dce_data_input; } pr_info("Got chunk process, status = 0x%x, sg_table[0].length = %d\n", @@ -537,7 +543,8 @@ static int do_test(struct dce_test_ctx *ctx, if (inf_process_req->dce_cf[0].length != input_len) { pr_err("Error %d\n", __LINE__); - return ret; + ret = -EINVAL; + goto fail_inflate_copy_input_to_dce_data; } print_dce_sg(inf_process_req->dce_cf[0]); @@ -546,7 +553,8 @@ static int do_test(struct dce_test_ctx *ctx, if (!inf_process_req->v_output) { pr_err("Error %d\n", __LINE__); - return ret; + ret = -ENOMEM; + goto fail_inflate_copy_input_to_dce_data; } inf_process_req->v_output_size = inf_process_req->dce_cf[0].length; @@ -555,14 +563,15 @@ static int do_test(struct dce_test_ctx *ctx, input_len); if (ret) { pr_err("Error %d\n", __LINE__); - return ret; + goto fail_inflate_copy_input_to_dce_data; } - /* compare output to orinal data */ + /* compare output to original data */ if (memcmp(inf_process_req->v_output, input_data, input_len)) { pr_err("Error %d\n", __LINE__); - return ret; + ret = -EINVAL; + goto fail_inflate_copy_input_to_dce_data; } - pr_info("Ouput inflate data matched original!\n"); + pr_info("Output inflate data matched original!\n"); /* Free dce data deflate operation, but keep vmalloc output */ free_dce_data(&inf_process_req->output_data); @@ -574,12 +583,13 @@ static int do_test(struct dce_test_ctx *ctx, kfree(def_process_req); ret = destroy_test_ctx(ctx); - if (ret) + if (ret) { pr_err("Error with test\n"); - else - pr_info("Done test loop\n"); + return ret; + } pr_info("Done test loop\n"); + return 0; fail_deflate_process: @@ -587,35 +597,70 @@ fail_deflate_process: &def_process_req->input_data, DMA_BIDIRECTIONAL); -fail_input_attach_data_to_sg_deflate: +fail_deflate_attach_data_list_to_sg_input: detach_data_list_from_sg(&def_process_req->dce_cf[0], &def_process_req->output_data, DMA_BIDIRECTIONAL); -fail_output_attach_data_to_sg_deflate: +fail_deflate_attach_data_list_to_sg_output: dma_unmap_dce_data(&def_process_req->output_data, DMA_BIDIRECTIONAL); -fail_dma_map_deflate_output_data: +fail_deflate_dma_map_dce_data_output: dma_unmap_dce_data(&def_process_req->input_data, DMA_BIDIRECTIONAL); -fail_alloc_dce_data_output: +fail_deflate_dma_map_dce_data_input: +fail_deflate_copy_input_to_dce_data: + free_dce_data(&def_process_req->output_data); + +fail_deflate_alloc_dce_data_output: free_dce_data(&def_process_req->input_data); -fail_alloc_dce_data_input: +fail_deflate_alloc_dce_data_input: vfree(def_process_req->v_output); -fail_v_output: +fail_deflate_v_output: kfree(def_process_req); -fail_inflate_params: - fsl_dce_chunk_destroy(&ctx->inflate_chunk, 0, NULL); - -fail_inflate_nop: -fail_deflate_nop: +fail_nop: destroy_test_ctx(ctx); fail_init_test_ctx: return ret; + +/* this section can't be added before deflate fail section +as it would cause seg fault */ +fail_inflate_process: + detach_data_list_from_sg(&inf_process_req->dce_cf[1], + &inf_process_req->input_data, + DMA_BIDIRECTIONAL); + +fail_inflate_attach_data_list_to_sg_input: + detach_data_list_from_sg(&inf_process_req->dce_cf[0], + &inf_process_req->output_data, + DMA_BIDIRECTIONAL); + +fail_inflate_attach_data_list_to_sg_output: + dma_unmap_dce_data(&inf_process_req->output_data, DMA_BIDIRECTIONAL); + +fail_inflate_dma_map_dce_data_output: + dma_unmap_dce_data(&inf_process_req->input_data, DMA_BIDIRECTIONAL); + +fail_inflate_dma_map_dce_data_input: +fail_inflate_copy_input_to_dce_data: + free_dce_data(&inf_process_req->output_data); + +fail_inflate_alloc_dce_data_output: + free_dce_data(&inf_process_req->input_data); + +fail_inflate_alloc_dce_data_input: + vfree(inf_process_req->v_output); + +fail_inflate_v_output: + kfree(inf_process_req); + +fail_inflate_params: + destroy_test_ctx(ctx); + return ret; } struct test_meta_info_t { diff --git a/drivers/staging/fsl_dce/tests/performance_simple/dce_perf_simple.c b/drivers/staging/fsl_dce/tests/performance_simple/dce_perf_simple.c index d91c42c..388f420 100644 --- a/drivers/staging/fsl_dce/tests/performance_simple/dce_perf_simple.c +++ b/drivers/staging/fsl_dce/tests/performance_simple/dce_perf_simple.c @@ -816,8 +816,10 @@ static int do_operation(void) pr_err("fsl_dce_chunk_deflate_params failed %d\n", ret); def_process_req = kzalloc(sizeof(*def_process_req), GFP_KERNEL); - if (!def_process_req) + if (!def_process_req) { pr_err("Line %d\n", __LINE__); + return -ENOMEM; + } init_completion(&def_process_req->cb_done); @@ -951,8 +953,10 @@ done: if (fsl_dce_get_status(def_process_req->output_fd.status) != STREAM_END) goto skip_output_copy; test_data->out_data = vmalloc(def_process_req->dce_cf[0].length); - if (!test_data->out_data) + if (!test_data->out_data) { pr_err("Unable to allocate output data\n"); + return -ENOMEM; + } test_data->out_data_len = def_process_req->dce_cf[0].length; if (!bman_output) { diff --git a/drivers/staging/fsl_dce/tests/performance_simple/dce_sf_perf_simple.c b/drivers/staging/fsl_dce/tests/performance_simple/dce_sf_perf_simple.c index d0e5fe0..10154b1 100644 --- a/drivers/staging/fsl_dce/tests/performance_simple/dce_sf_perf_simple.c +++ b/drivers/staging/fsl_dce/tests/performance_simple/dce_sf_perf_simple.c @@ -926,8 +926,10 @@ static int do_operation(void) for (i = 0; i < chunk_count; i++) { def_process_req = kzalloc(sizeof(*def_process_req), GFP_KERNEL); - if (!def_process_req) + if (!def_process_req) { pr_err("Line %d\n", __LINE__); + return -ENOMEM; + } def_process_req->extra_data_size = i; @@ -1065,6 +1067,11 @@ try_again: } i++; } + + if (!def_process_req) { + pr_err("Line %d\n", __LINE__); + return -EINVAL; + } /* wait for last request to be processed */ wait_for_completion(&def_process_req->cb_done); end_time = mfatb(); @@ -1092,8 +1099,10 @@ done: pr_info("Total output required %d\n", total_out); test_data->out_data_len = total_out; test_data->out_data = vmalloc(total_out); - if (!test_data->out_data) + if (!test_data->out_data) { pr_err("vmalloc FAILED\n"); + return -ENOMEM; + } p_out = test_data->out_data; /* copy output */ diff --git a/drivers/staging/fsl_pme2/pme2_ctrl.c b/drivers/staging/fsl_pme2/pme2_ctrl.c index 388162b..56e2443 100644 --- a/drivers/staging/fsl_pme2/pme2_ctrl.c +++ b/drivers/staging/fsl_pme2/pme2_ctrl.c @@ -138,6 +138,10 @@ static __init int parse_mem_property(struct device_node *node, const char *name, } else if (zero) { /* map as cacheable, non-guarded */ void *tmpp = ioremap_prot(*addr, *sz, 0); + if (!tmpp) { + pr_err("pme: Failed to remap\n"); + return -EINVAL; + } memset(tmpp, 0, *sz); iounmap(tmpp); } diff --git a/drivers/staging/fsl_pme2/pme2_db.c b/drivers/staging/fsl_pme2/pme2_db.c index 50263d5..0cc13fc 100644 --- a/drivers/staging/fsl_pme2/pme2_db.c +++ b/drivers/staging/fsl_pme2/pme2_db.c @@ -295,6 +295,7 @@ comp_frame_free_rx: kfree(rx_data); goto unmap_input_frame; single_frame_unmap_frame: + kfree(rx_data); unmap_input_frame: free_tx_data: kfree(tx_data); diff --git a/drivers/staging/fsl_pme2/pme2_high.c b/drivers/staging/fsl_pme2/pme2_high.c index 12ab7a1..e39bfc6 100644 --- a/drivers/staging/fsl_pme2/pme2_high.c +++ b/drivers/staging/fsl_pme2/pme2_high.c @@ -933,6 +933,7 @@ static inline void cb_helper(__always_unused struct qman_portal *portal, if (error) do_flags(ctx, 0, 0, PME_CTX_FLAG_DEAD, 0); token = pop_matching_token(ctx, fd); + BUG_ON(!token); if (likely(token->cmd_type == pme_cmd_scan)) { ctx->cb(ctx, fd, token); } else if (token->cmd_type == pme_cmd_pmtcc) { @@ -1002,6 +1003,7 @@ static void cb_ern(__always_unused struct qman_portal *portal, ctx = data->parent; token = pop_matching_token(ctx, &mr->ern.fd); + BUG_ON(!token); if (likely(token->cmd_type == pme_cmd_scan)) { BUG_ON(!ctx->ern_cb); ctx->ern_cb(ctx, mr, token); diff --git a/drivers/staging/fsl_qbman/qman_high.c b/drivers/staging/fsl_qbman/qman_high.c index d17d301..80604b7 100644 --- a/drivers/staging/fsl_qbman/qman_high.c +++ b/drivers/staging/fsl_qbman/qman_high.c @@ -3397,7 +3397,7 @@ int qman_ceetm_sp_release(struct qm_ceetm_sp *sp) { struct qm_ceetm_sp *p; - if (sp->lni->is_claimed == 1) { + if (sp->lni && sp->lni->is_claimed == 1) { pr_err("The dependency of sub-portal has not been released!\n"); return -EBUSY; } @@ -4456,7 +4456,9 @@ int qman_ceetm_cq_claim(struct qm_ceetm_cq **cq, if (qman_ceetm_configure_cq(&cq_config)) { pr_err("Can't configure the CQ#%d with CCGRID#%d\n", idx, ccg->idx); - return -EINVAL; + list_del(&p->node); + kfree(p); + return -EINVAL; } } @@ -4503,6 +4505,8 @@ int qman_ceetm_cq_claim_A(struct qm_ceetm_cq **cq, if (qman_ceetm_configure_cq(&cq_config)) { pr_err("Can't configure the CQ#%d with CCGRID#%d\n", idx, ccg->idx); + list_del(&p->node); + kfree(p); return -EINVAL; } } @@ -4549,6 +4553,8 @@ int qman_ceetm_cq_claim_B(struct qm_ceetm_cq **cq, if (qman_ceetm_configure_cq(&cq_config)) { pr_err("Can't configure the CQ#%d with CCGRID#%d\n", idx, ccg->idx); + list_del(&p->node); + kfree(p); return -EINVAL; } } @@ -4814,6 +4820,8 @@ int qman_ceetm_lfq_claim(struct qm_ceetm_lfq **lfq, if (qman_ceetm_configure_lfqmt(&lfqmt_config)) { pr_err("Can't configure LFQMT for LFQID#%d @ CQ#%d\n", lfqid, cq->idx); + list_del(&p->node); + kfree(p); return -EINVAL; } *lfq = p; diff --git a/net/core/dev.c b/net/core/dev.c index dd354f0..71d316a 100644 --- a/net/core/dev.c +++ b/net/core/dev.c @@ -2470,7 +2470,7 @@ static void dev_gso_skb_destructor(struct sk_buff *skb) * This function segments the given skb and stores the list of segments * in skb->next. */ -static int dev_gso_segment(struct sk_buff *skb, netdev_features_t features) +int dev_gso_segment(struct sk_buff *skb, netdev_features_t features) { struct sk_buff *segs; @@ -2489,6 +2489,13 @@ static int dev_gso_segment(struct sk_buff *skb, netdev_features_t features) return 0; } +EXPORT_SYMBOL(dev_gso_segment); + +void dev_set_skb_destructor(struct sk_buff *skb) +{ + skb->destructor = DEV_GSO_CB(skb)->destructor; +} +EXPORT_SYMBOL(dev_set_skb_destructor); static netdev_features_t harmonize_features(struct sk_buff *skb, const struct net_device *dev, |