diff --git a/ompi/mca/osc/rdma/osc_rdma_active_target.c b/ompi/mca/osc/rdma/osc_rdma_active_target.c index 0275f272441..c3c52aba478 100644 --- a/ompi/mca/osc/rdma/osc_rdma_active_target.c +++ b/ompi/mca/osc/rdma/osc_rdma_active_target.c @@ -385,6 +385,8 @@ int ompi_osc_rdma_start_atomic (ompi_group_t *group, int mpi_assert, ompi_win_t OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_TRACE, "start group size %d", sync->num_peers); + sync->type = OMPI_OSC_RDMA_SYNC_TYPE_PSCW; + if (0 == ompi_group_size (group)) { /* nothing more to do. this is an empty start epoch */ OPAL_THREAD_UNLOCK(&module->lock); @@ -393,8 +395,6 @@ int ompi_osc_rdma_start_atomic (ompi_group_t *group, int mpi_assert, ompi_win_t opal_atomic_wmb (); - sync->type = OMPI_OSC_RDMA_SYNC_TYPE_PSCW; - /* prevent us from entering a passive-target, fence, or another pscw access epoch until * the matching complete is called */ sync->epoch_active = true; @@ -466,17 +466,19 @@ int ompi_osc_rdma_complete_atomic (ompi_win_t *win) sync->type = OMPI_OSC_RDMA_SYNC_TYPE_NONE; sync->epoch_active = false; - /* phase 2 cleanup group */ - OBJ_RELEASE(group); - peers = sync->peer_list.peers; if (NULL == peers) { /* empty peer list */ OPAL_THREAD_UNLOCK(&(module->lock)); - OBJ_RELEASE(group); + if(MPI_GROUP_EMPTY != group) { + OBJ_RELEASE(group); + } return OMPI_SUCCESS; } + /* phase 2 cleanup group */ + OBJ_RELEASE(group); + sync->peer_list.peers = NULL; OPAL_THREAD_UNLOCK(&(module->lock)); diff --git a/ompi/mca/osc/rdma/osc_rdma_component.c b/ompi/mca/osc/rdma/osc_rdma_component.c index 74398060081..24a4bd4f549 100644 --- a/ompi/mca/osc/rdma/osc_rdma_component.c +++ b/ompi/mca/osc/rdma/osc_rdma_component.c @@ -521,7 +521,7 @@ static int allocate_state_single (ompi_osc_rdma_module_t *module, void **base, s my_peer->flags |= OMPI_OSC_RDMA_PEER_LOCAL_BASE; my_peer->state = (uint64_t) (uintptr_t) module->state; - if (module->use_cpu_atomics) { + if (module->use_cpu_atomics || my_peer->rank == my_rank) { /* all peers are local or it is safe to mix cpu and nic atomics */ my_peer->flags |= OMPI_OSC_RDMA_PEER_LOCAL_STATE; } else { @@ -581,7 +581,7 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s int my_rank = ompi_comm_rank (module->comm); int global_size = ompi_comm_size (module->comm); ompi_osc_rdma_region_t *state_region; - struct _local_data *temp; + struct _local_data *temp = NULL; char *data_file; int page_size = opal_getpagesize(); @@ -596,7 +596,7 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s if (!module->single_node) { for (int i = 0 ; i < module->btls_in_use ; ++i) { - module->use_cpu_atomics = module->use_cpu_atomics && !!(module->selected_btls[i]->btl_flags & MCA_BTL_ATOMIC_SUPPORTS_GLOB); + module->use_cpu_atomics = module->use_cpu_atomics && !!(module->selected_btls[i]->btl_atomic_flags & MCA_BTL_ATOMIC_SUPPORTS_GLOB); } } @@ -624,13 +624,12 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s size += OPAL_ALIGN_PAD_AMOUNT(size, page_size); } - do { - temp = calloc (local_size, sizeof (temp[0])); - if (NULL == temp) { - ret = OMPI_ERR_OUT_OF_RESOURCE; - break; - } + temp = calloc (local_size, sizeof (temp[0])); + if (NULL == temp) { + return OMPI_ERR_OUT_OF_RESOURCE; + } + do { temp[local_rank].rank = my_rank; temp[local_rank].size = size; @@ -777,7 +776,7 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s ex_peer = (ompi_osc_rdma_peer_extended_t *) peer; /* set up peer state */ - if (module->use_cpu_atomics) { + if (module->use_cpu_atomics || peer->rank == my_rank) { /* all peers are local or it is safe to mix cpu and nic atomics */ peer->flags |= OMPI_OSC_RDMA_PEER_LOCAL_STATE; peer->state = (osc_rdma_counter_t) peer_state; @@ -788,10 +787,8 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s peer->state_handle = (mca_btl_base_registration_handle_t *) state_region->btl_handle_data; } peer->state = (osc_rdma_counter_t) ((uintptr_t) state_region->base + state_base + module->state_size * i); - if (i > 0) { - peer->state_endpoint = local_leader->state_endpoint; - peer->state_btl_index = local_leader->state_btl_index; - } + peer->state_endpoint = local_leader->data_endpoint; // data_endpoint initialized in ompi_osc_rdma_new_peer(); + peer->state_btl_index = local_leader->data_btl_index; } if (my_rank == peer_rank) { @@ -808,7 +805,7 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s ompi_osc_module_add_peer (module, peer); if (MPI_WIN_FLAVOR_DYNAMIC == module->flavor) { - if (module->use_cpu_atomics && peer_rank == my_rank) { + if (peer_rank == my_rank) { peer->flags |= OMPI_OSC_RDMA_PEER_LOCAL_BASE; } /* nothing more to do */ @@ -824,7 +821,7 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s ex_peer->size = temp[i].size; } - if (module->use_cpu_atomics && (MPI_WIN_FLAVOR_ALLOCATE == module->flavor || peer_rank == my_rank)) { + if (module->use_cpu_atomics && (MPI_WIN_FLAVOR_ALLOCATE == module->flavor)) { /* base is local and cpu atomics are available */ if (MPI_WIN_FLAVOR_ALLOCATE == module->flavor) { ex_peer->super.base = (uintptr_t) module->segment_base + offset; @@ -832,7 +829,6 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s ex_peer->super.base = (uintptr_t) *base; } - peer->flags |= OMPI_OSC_RDMA_PEER_LOCAL_BASE; offset += temp[i].size; } else { ex_peer->super.base = peer_region->base; @@ -841,6 +837,10 @@ static int allocate_state_shared (ompi_osc_rdma_module_t *module, void **base, s ex_peer->super.base_handle = (mca_btl_base_registration_handle_t *) peer_region->btl_handle_data; } } + + if(my_rank == peer_rank) { + peer->flags |= OMPI_OSC_RDMA_PEER_LOCAL_BASE; + } } } while (0); @@ -914,10 +914,8 @@ static void ompi_osc_rdma_ensure_local_add_procs (void) static int ompi_osc_rdma_query_alternate_btls (ompi_communicator_t *comm, ompi_osc_rdma_module_t *module) { mca_btl_base_selected_module_t *item; - char **btls_to_use = opal_argv_split (ompi_osc_rdma_btl_alternate_names, ','); int btls_found = 0; - - btls_to_use = opal_argv_split (ompi_osc_rdma_btl_alternate_names, ','); + char **btls_to_use = opal_argv_split (ompi_osc_rdma_btl_alternate_names, ','); if (NULL == btls_to_use) { OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_INFO, "no alternate BTLs requested: %s", ompi_osc_rdma_btl_alternate_names); return OMPI_ERR_UNREACH; diff --git a/ompi/mca/osc/rdma/osc_rdma_dynamic.c b/ompi/mca/osc/rdma/osc_rdma_dynamic.c index 8adfa7f8159..628dcd7789a 100644 --- a/ompi/mca/osc/rdma/osc_rdma_dynamic.c +++ b/ompi/mca/osc/rdma/osc_rdma_dynamic.c @@ -428,10 +428,6 @@ static int ompi_osc_rdma_refresh_dynamic_region (ompi_osc_rdma_module_t *module, } peer->regions = temp; - /* lock the region */ - ompi_osc_rdma_lock_acquire_shared (module, &peer->super, 1, offsetof (ompi_osc_rdma_state_t, regions_lock), - OMPI_OSC_RDMA_LOCK_EXCLUSIVE); - source_address = (uint64_t)(intptr_t) peer->super.state + offsetof (ompi_osc_rdma_state_t, regions); ret = ompi_osc_get_data_blocking (module, peer->super.state_btl_index, peer->super.state_endpoint, source_address, peer->super.state_handle, peer->regions, region_len); @@ -440,9 +436,6 @@ static int ompi_osc_rdma_refresh_dynamic_region (ompi_osc_rdma_module_t *module, return ret; } - /* release the region lock */ - ompi_osc_rdma_lock_release_shared (module, &peer->super, -1, offsetof (ompi_osc_rdma_state_t, regions_lock)); - /* update cached region ids */ peer->region_id = region_id; peer->region_count = region_count; diff --git a/ompi/mca/osc/rdma/osc_rdma_lock.h b/ompi/mca/osc/rdma/osc_rdma_lock.h index 36a30a1cc0b..74aed2770a0 100644 --- a/ompi/mca/osc/rdma/osc_rdma_lock.h +++ b/ompi/mca/osc/rdma/osc_rdma_lock.h @@ -86,10 +86,12 @@ static inline int ompi_osc_rdma_btl_fop (ompi_osc_rdma_module_t *module, uint8_t if (OPAL_SUCCESS != ret) { if (OPAL_LIKELY(1 == ret)) { - *result = ((int64_t *) pending_op->op_buffer)[0]; ret = OMPI_SUCCESS; - ompi_osc_rdma_atomic_complete (selected_btl, endpoint, pending_op->op_buffer, - pending_op->op_frag->handle, (void *) pending_op, NULL, OPAL_SUCCESS); + if(false == pending_op->op_complete) { + *result = ((int64_t *) pending_op->op_buffer)[0]; + ompi_osc_rdma_atomic_complete (selected_btl, endpoint, pending_op->op_buffer, + pending_op->op_frag->handle, (void *) pending_op, NULL, OPAL_SUCCESS); + } } else { /* need to release here because ompi_osc_rdma_atomic_complete was not called */ OBJ_RELEASE(pending_op); @@ -161,8 +163,10 @@ static inline int ompi_osc_rdma_btl_op (ompi_osc_rdma_module_t *module, uint8_t /* need to release here because ompi_osc_rdma_atomic_complete was not called */ OBJ_RELEASE(pending_op); if (OPAL_LIKELY(1 == ret)) { - if (cbfunc) { - cbfunc (cbdata, cbcontext, OMPI_SUCCESS); + if(false == pending_op->op_complete) { + if (cbfunc) { + cbfunc (cbdata, cbcontext, OMPI_SUCCESS); + } } ret = OMPI_SUCCESS; } diff --git a/ompi/mca/osc/rdma/osc_rdma_peer.c b/ompi/mca/osc/rdma/osc_rdma_peer.c index c6689d78812..c67a3a29e52 100644 --- a/ompi/mca/osc/rdma/osc_rdma_peer.c +++ b/ompi/mca/osc/rdma/osc_rdma_peer.c @@ -71,6 +71,7 @@ static int ompi_osc_rdma_peer_btl_endpoint (struct ompi_osc_rdma_module_t *modul } } + OSC_RDMA_VERBOSE(MCA_BASE_VERBOSE_DEBUG, "WARNING - couldn't find a connection"); /* unlikely but can happen when creating a peer for self */ return OMPI_ERR_UNREACH; } diff --git a/opal/mca/btl/base/btl_base_am_rdma.c b/opal/mca/btl/base/btl_base_am_rdma.c index 81a0d74b266..3f17bd1d701 100644 --- a/opal/mca/btl/base/btl_base_am_rdma.c +++ b/opal/mca/btl/base/btl_base_am_rdma.c @@ -614,7 +614,7 @@ static int mca_btl_base_am_rdma_respond(mca_btl_base_module_t *btl, send_descriptor->des_cbfunc = NULL; int ret = btl->btl_send(btl, endpoint, send_descriptor, mca_btl_base_rdma_resp_tag()); - if (OPAL_UNLIKELY(OPAL_SUCCESS != ret)) { + if (OPAL_UNLIKELY(ret < 0)) { *descriptor = send_descriptor; } return ret; @@ -635,7 +635,7 @@ mca_btl_base_am_rmda_rdma_complete(mca_btl_base_module_t *btl, operation->is_completed = true; int ret = mca_btl_base_am_rdma_respond(operation->btl, operation->endpoint, &operation->descriptor, NULL, &operation->hdr); - if (OPAL_UNLIKELY(OPAL_SUCCESS != ret)) { + if (OPAL_UNLIKELY(ret < 0)) { BTL_VERBOSE( ("could not send a response. queueing the response for later. endpoint=%p, ret=%d", endpoint, ret)); @@ -781,7 +781,7 @@ static int mca_btl_base_am_rdma_progress(void) int ret = descriptor->btl->btl_send(descriptor->btl, descriptor->endpoint, descriptor->descriptor, mca_btl_base_rdma_tag(context->type)); - if (OPAL_SUCCESS == ret) { + if (ret <= 1) { opal_list_remove_item(&default_module.queued_initiator_descriptors, &descriptor->super); } @@ -932,7 +932,7 @@ static void mca_btl_base_am_process_rdma(mca_btl_base_module_t *btl, abort(); } - if (OPAL_SUCCESS != ret) { + if (ret < 0) { mca_btl_base_rdma_queue_operation(btl, desc->endpoint, descriptor, 0, hdr, operation); } } @@ -995,7 +995,7 @@ static void mca_btl_base_am_process_atomic(mca_btl_base_module_t *btl, mca_btl_base_descriptor_t *descriptor = NULL; int ret = mca_btl_base_am_rdma_respond(btl, desc->endpoint, &descriptor, &atomic_response, hdr); - if (OPAL_SUCCESS != ret) { + if (ret < 0) { mca_btl_base_rdma_queue_operation(btl, desc->endpoint, descriptor, atomic_response, hdr, NULL); } diff --git a/opal/mca/btl/tcp/btl_tcp.c b/opal/mca/btl/tcp/btl_tcp.c index 4fe37f566bb..9769fd36198 100644 --- a/opal/mca/btl/tcp/btl_tcp.c +++ b/opal/mca/btl/tcp/btl_tcp.c @@ -54,7 +54,6 @@ mca_btl_tcp_module_t mca_btl_tcp_module = .btl_free = mca_btl_tcp_free, .btl_prepare_src = mca_btl_tcp_prepare_src, .btl_send = mca_btl_tcp_send, - .btl_put = mca_btl_tcp_put, .btl_dump = mca_btl_base_dump, .btl_register_error = mca_btl_tcp_register_error_cb, /* register error */ }, @@ -330,151 +329,6 @@ int mca_btl_tcp_send(struct mca_btl_base_module_t *btl, struct mca_btl_base_endp return mca_btl_tcp_endpoint_send(endpoint, frag); } -static void fake_rdma_complete(mca_btl_base_module_t *btl, mca_btl_base_endpoint_t *endpoint, - mca_btl_base_descriptor_t *desc, int rc) -{ - mca_btl_tcp_frag_t *frag = (mca_btl_tcp_frag_t *) desc; - - frag->cb.func(btl, endpoint, frag->segments[0].seg_addr.pval, NULL, frag->cb.context, - frag->cb.data, rc); -} - -/** - * Initiate an asynchronous put. - */ - -int mca_btl_tcp_put(mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *endpoint, - void *local_address, uint64_t remote_address, - mca_btl_base_registration_handle_t *local_handle, - mca_btl_base_registration_handle_t *remote_handle, size_t size, int flags, - int order, mca_btl_base_rdma_completion_fn_t cbfunc, void *cbcontext, - void *cbdata) -{ - mca_btl_tcp_module_t *tcp_btl = (mca_btl_tcp_module_t *) btl; - mca_btl_tcp_frag_t *frag = NULL; - int i; - - MCA_BTL_TCP_FRAG_ALLOC_USER(frag); - if (OPAL_UNLIKELY(NULL == frag)) { - return OPAL_ERR_OUT_OF_RESOURCE; - } - - frag->endpoint = endpoint; - - frag->segments->seg_len = size; - frag->segments->seg_addr.pval = local_address; - - frag->base.des_segments = frag->segments; - frag->base.des_segment_count = 1; - frag->base.order = MCA_BTL_NO_ORDER; - - frag->segments[0].seg_addr.pval = local_address; - frag->segments[0].seg_len = size; - - frag->segments[1].seg_addr.lval = remote_address; - frag->segments[1].seg_len = size; - if (endpoint->endpoint_nbo) { - MCA_BTL_BASE_SEGMENT_HTON(frag->segments[1]); - } - - frag->base.des_flags = MCA_BTL_DES_FLAGS_BTL_OWNERSHIP | MCA_BTL_DES_SEND_ALWAYS_CALLBACK; - frag->base.des_cbfunc = fake_rdma_complete; - - frag->cb.func = cbfunc; - frag->cb.data = cbdata; - frag->cb.context = cbcontext; - - frag->btl = tcp_btl; - frag->endpoint = endpoint; - frag->rc = 0; - frag->iov_idx = 0; - frag->hdr.size = 0; - frag->iov_cnt = 2; - frag->iov_ptr = frag->iov; - frag->iov[0].iov_base = (IOVBASE_TYPE *) &frag->hdr; - frag->iov[0].iov_len = sizeof(frag->hdr); - frag->iov[1].iov_base = (IOVBASE_TYPE *) (frag->segments + 1); - frag->iov[1].iov_len = sizeof(mca_btl_base_segment_t); - for (i = 0; i < (int) frag->base.des_segment_count; i++) { - frag->hdr.size += frag->segments[i].seg_len; - frag->iov[i + 2].iov_len = frag->segments[i].seg_len; - frag->iov[i + 2].iov_base = (IOVBASE_TYPE *) frag->segments[i].seg_addr.pval; - frag->iov_cnt++; - } - frag->hdr.base.tag = MCA_BTL_TAG_BTL; - frag->hdr.type = MCA_BTL_TCP_HDR_TYPE_PUT; - frag->hdr.count = 1; - if (endpoint->endpoint_nbo) { - MCA_BTL_TCP_HDR_HTON(frag->hdr); - } - return ((i = mca_btl_tcp_endpoint_send(endpoint, frag)) >= 0 ? OPAL_SUCCESS : i); -} - -/** - * Initiate an asynchronous get. - */ - -int mca_btl_tcp_get(mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *endpoint, - void *local_address, uint64_t remote_address, - mca_btl_base_registration_handle_t *local_handle, - mca_btl_base_registration_handle_t *remote_handle, size_t size, int flags, - int order, mca_btl_base_rdma_completion_fn_t cbfunc, void *cbcontext, - void *cbdata) -{ - mca_btl_tcp_module_t *tcp_btl = (mca_btl_tcp_module_t *) btl; - mca_btl_tcp_frag_t *frag = NULL; - int rc; - - MCA_BTL_TCP_FRAG_ALLOC_USER(frag); - if (OPAL_UNLIKELY(NULL == frag)) { - return OPAL_ERR_OUT_OF_RESOURCE; - ; - } - - frag->endpoint = endpoint; - - frag->segments->seg_len = size; - frag->segments->seg_addr.pval = local_address; - - frag->base.des_segments = frag->segments; - frag->base.des_segment_count = 1; - frag->base.order = MCA_BTL_NO_ORDER; - - frag->segments[0].seg_addr.pval = local_address; - frag->segments[0].seg_len = size; - - frag->segments[1].seg_addr.lval = remote_address; - frag->segments[1].seg_len = size; - - /* call the rdma callback through the descriptor callback. this is - * tcp so the extra latency is not an issue */ - frag->base.des_flags = MCA_BTL_DES_FLAGS_BTL_OWNERSHIP | MCA_BTL_DES_SEND_ALWAYS_CALLBACK; - frag->base.des_cbfunc = fake_rdma_complete; - - frag->cb.func = cbfunc; - frag->cb.data = cbdata; - frag->cb.context = cbcontext; - - frag->btl = tcp_btl; - frag->endpoint = endpoint; - frag->rc = 0; - frag->iov_idx = 0; - frag->hdr.size = 0; - frag->iov_cnt = 2; - frag->iov_ptr = frag->iov; - frag->iov[0].iov_base = (IOVBASE_TYPE *) &frag->hdr; - frag->iov[0].iov_len = sizeof(frag->hdr); - frag->iov[1].iov_base = (IOVBASE_TYPE *) &frag->segments[1]; - frag->iov[1].iov_len = sizeof(mca_btl_base_segment_t); - frag->hdr.base.tag = MCA_BTL_TAG_BTL; - frag->hdr.type = MCA_BTL_TCP_HDR_TYPE_GET; - frag->hdr.count = 1; - if (endpoint->endpoint_nbo) { - MCA_BTL_TCP_HDR_HTON(frag->hdr); - } - return ((rc = mca_btl_tcp_endpoint_send(endpoint, frag)) >= 0 ? OPAL_SUCCESS : rc); -} - /* * Cleanup/release module resources. */ diff --git a/opal/mca/btl/tcp/btl_tcp.h b/opal/mca/btl/tcp/btl_tcp.h index 55fed24398b..07c602cfdd7 100644 --- a/opal/mca/btl/tcp/btl_tcp.h +++ b/opal/mca/btl/tcp/btl_tcp.h @@ -251,28 +251,6 @@ extern int mca_btl_tcp_send(struct mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *btl_peer, struct mca_btl_base_descriptor_t *descriptor, mca_btl_base_tag_t tag); -/** - * Initiate an asynchronous put. - */ - -int mca_btl_tcp_put(mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *endpoint, - void *local_address, uint64_t remote_address, - mca_btl_base_registration_handle_t *local_handle, - mca_btl_base_registration_handle_t *remote_handle, size_t size, int flags, - int order, mca_btl_base_rdma_completion_fn_t cbfunc, void *cbcontext, - void *cbdata); - -/** - * Initiate an asynchronous get. - */ - -int mca_btl_tcp_get(mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *endpoint, - void *local_address, uint64_t remote_address, - mca_btl_base_registration_handle_t *local_handle, - mca_btl_base_registration_handle_t *remote_handle, size_t size, int flags, - int order, mca_btl_base_rdma_completion_fn_t cbfunc, void *cbcontext, - void *cbdata); - /** * Allocate a descriptor with a segment of the requested size. * Note that the BTL layer may choose to return a smaller size diff --git a/opal/mca/btl/tcp/btl_tcp_frag.c b/opal/mca/btl/tcp/btl_tcp_frag.c index e401a48b81e..f15ab1d7ca3 100644 --- a/opal/mca/btl/tcp/btl_tcp_frag.c +++ b/opal/mca/btl/tcp/btl_tcp_frag.c @@ -288,25 +288,6 @@ bool mca_btl_tcp_frag_recv(mca_btl_tcp_frag_t *frag, int sd) goto repeat; } break; - case MCA_BTL_TCP_HDR_TYPE_PUT: - if (frag->iov_idx == 1) { - frag->iov[1].iov_base = (IOVBASE_TYPE *) frag->segments; - frag->iov[1].iov_len = frag->hdr.count * sizeof(mca_btl_base_segment_t); - frag->iov_cnt++; - goto repeat; - } else if (frag->iov_idx == 2) { - for (i = 0; i < frag->hdr.count; i++) { - if (btl_endpoint->endpoint_nbo) { - MCA_BTL_BASE_SEGMENT_NTOH(frag->segments[i]); - } - frag->iov[i + 2].iov_base = (IOVBASE_TYPE *) frag->segments[i].seg_addr.pval; - frag->iov[i + 2].iov_len = frag->segments[i].seg_len; - } - frag->iov_cnt += frag->hdr.count; - goto repeat; - } - break; - case MCA_BTL_TCP_HDR_TYPE_GET: default: break; } diff --git a/opal/mca/btl/tcp/btl_tcp_hdr.h b/opal/mca/btl/tcp/btl_tcp_hdr.h index 69ec22f6fdf..f7d663a7e43 100644 --- a/opal/mca/btl/tcp/btl_tcp_hdr.h +++ b/opal/mca/btl/tcp/btl_tcp_hdr.h @@ -30,8 +30,6 @@ BEGIN_C_DECLS */ #define MCA_BTL_TCP_HDR_TYPE_SEND 1 -#define MCA_BTL_TCP_HDR_TYPE_PUT 2 -#define MCA_BTL_TCP_HDR_TYPE_GET 3 #define MCA_BTL_TCP_HDR_TYPE_FIN 4 /* The MCA_BTL_TCP_HDR_TYPE_FIN is a special kind of message sent during normal * connexion closing. Before the endpoint closes the socket, it performs a