summaryrefslogtreecommitdiffstats
path: root/rpc/rpc-transport/rdma/src/rdma.c
diff options
context:
space:
mode:
authorYi Wang <wangyi@storswift.com>2017-12-26 01:17:04 +0800
committerAmar Tumballi <amarts@redhat.com>2017-12-28 06:59:04 +0000
commit3f0405f3ee81e05666116720c4f136403f778188 (patch)
treed245abca5798227e16217466192a6f2613bb7a2b /rpc/rpc-transport/rdma/src/rdma.c
parent98bbc3c94244b6bc9fd18b698ad26e2c867ca63e (diff)
rpc-transport/rdma: Fix coverity issues in rdma transport
Issues: [1] https://download.gluster.org/pub/gluster/glusterfs/static-analysis/master/glusterfs-coverity/2017-12-25-0bc22bef/cov-errors.txt Resolved: [1] rdma.c:128 var_deref_op: Dereferencing null pointer "post". [2] rdma.c:677: Potentially overflowing expression [3] rdma.c:4250: freed_arg: "rpc_transport_unref" frees "peer->trans". [4] rdma.c:4644: var_deref_op: Dereferencing null pointer "rdma_ctx". [5] rdma.c:4945: cond_false: Condition "rdma_ctx != NULL", taking false branch. Change-Id: Iec38f118d645df4131739da412a6c741ebbd2f85 BUG: 789278 Signed-off-by: Yi Wang <wangyi@storswift.com>
Diffstat (limited to 'rpc/rpc-transport/rdma/src/rdma.c')
-rw-r--r--rpc/rpc-transport/rdma/src/rdma.c66
1 files changed, 34 insertions, 32 deletions
diff --git a/rpc/rpc-transport/rdma/src/rdma.c b/rpc/rpc-transport/rdma/src/rdma.c
index f034ed1..1595be2 100644
--- a/rpc/rpc-transport/rdma/src/rdma.c
+++ b/rpc/rpc-transport/rdma/src/rdma.c
@@ -124,7 +124,7 @@ gf_rdma_new_post (rpc_transport_t *this, gf_rdma_device_t *device, int32_t len,
ret = 0;
out:
- if (ret != 0) {
+ if (ret != 0 && post) {
free (post->buf);
GF_FREE (post);
@@ -674,7 +674,7 @@ gf_rdma_create_cq (rpc_transport_t *this)
goto out;
}
- send_cqe = options->send_count * 128;
+ send_cqe = (uint64_t)options->send_count * 128;
send_cqe = (send_cqe > device_attr.max_cqe)
? device_attr.max_cqe : send_cqe;
@@ -4224,38 +4224,39 @@ gf_rdma_recv_completion_proc (void *data)
peer = __gf_rdma_lookup_peer (device,
wc[index].qp_num);
- /*
- * keep a refcount on transport so that it
- * does not get freed because of some error
- * indicated by wc.status till we are done
- * with usage of peer and thereby that of
- * trans.
- */
- if (peer != NULL) {
- rpc_transport_ref (peer->trans);
- }
+ /*
+ * keep a refcount on transport so that it
+ * does not get freed because of some error
+ * indicated by wc.status till we are done
+ * with usage of peer and thereby that of
+ * trans.
+ */
+ if (peer != NULL) {
+ rpc_transport_ref (peer->trans);
+ }
}
pthread_mutex_unlock (&device->qpreg.lock);
if (wc[index].status != IBV_WC_SUCCESS) {
gf_msg (GF_RDMA_LOG_NAME,
- GF_LOG_ERROR, 0,
- RDMA_MSG_RECV_ERROR, "recv work "
- "request on `%s' returned error (%d)",
- device->device_name,
- wc[index].status);
- failed = 1;
- if (peer) {
- ibv_ack_cq_events (event_cq, num_wr);
- rpc_transport_unref (peer->trans);
- rpc_transport_disconnect (peer->trans,
- _gf_false);
- }
+ GF_LOG_ERROR, 0,
+ RDMA_MSG_RECV_ERROR, "recv work "
+ "request on `%s' returned error (%d)",
+ device->device_name,
+ wc[index].status);
+ failed = 1;
+ if (peer) {
+ ibv_ack_cq_events (event_cq, num_wr);
+ rpc_transport_disconnect (peer->trans,
+ _gf_false);
+ rpc_transport_unref (peer->trans);
+ }
- if (post) {
- gf_rdma_post_unref (post);
- }
- continue;
+ if (post) {
+ gf_rdma_post_unref (post);
+ }
+
+ continue;
}
if (peer) {
@@ -4640,7 +4641,7 @@ __gf_rdma_ctx_create (void)
}
out:
- if (ret < 0) {
+ if (ret < 0 && rdma_ctx) {
if (rdma_ctx->rdma_cm_event_channel != NULL) {
rdma_destroy_event_channel (rdma_ctx->rdma_cm_event_channel);
}
@@ -4940,17 +4941,18 @@ init (rpc_transport_t *this)
return -1;
}
rdma_ctx = this->ctx->ib;
+ if (!rdma_ctx)
+ return -1;
+
pthread_mutex_lock (&rdma_ctx->lock);
{
- if (rdma_ctx != NULL) {
- if (this->dl_handle && (++(rdma_ctx->dlcount)) == 1) {
+ if (this->dl_handle && (++(rdma_ctx->dlcount)) == 1) {
iobuf_pool = this->ctx->iobuf_pool;
iobuf_pool->rdma_registration = gf_rdma_register_arena;
iobuf_pool->rdma_deregistration =
gf_rdma_deregister_arena;
gf_rdma_register_iobuf_pool_with_device
(rdma_ctx->device, iobuf_pool);
- }
}
}
pthread_mutex_unlock (&rdma_ctx->lock);