X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/385f6d5ce3f2aff9707045601b32a66b017d7bed..98a4fb37cf38be98161ec6fc43ed0733eaefcbee:/src/smpi/bindings/smpi_pmpi_request.cpp diff --git a/src/smpi/bindings/smpi_pmpi_request.cpp b/src/smpi/bindings/smpi_pmpi_request.cpp index d89bddec70..f64dbabda6 100644 --- a/src/smpi/bindings/smpi_pmpi_request.cpp +++ b/src/smpi/bindings/smpi_pmpi_request.cpp @@ -66,6 +66,12 @@ int PMPI_Recv_init(void *buf, int count, MPI_Datatype datatype, int src, int tag return retval; } +int PMPI_Rsend_init(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm, + MPI_Request* request) +{ + return PMPI_Send_init(buf, count, datatype, dst, tag, comm, request); +} + int PMPI_Ssend_init(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm, MPI_Request* request) { int retval = 0; @@ -139,10 +145,9 @@ int PMPI_Startall(int count, MPI_Request * requests) if(retval != MPI_ERR_REQUEST) { int my_proc_id = simgrid::s4u::this_actor::get_pid(); TRACE_smpi_comm_in(my_proc_id, __func__, new simgrid::instr::NoOpTIData("Startall")); - MPI_Request req = MPI_REQUEST_NULL; if (not TRACE_smpi_view_internals()) for (int i = 0; i < count; i++) { - req = requests[i]; + MPI_Request req = requests[i]; if (req->flags() & MPI_REQ_SEND) TRACE_smpi_send(my_proc_id, my_proc_id, getPid(req->comm(), req->dst()), req->tag(), req->size()); } @@ -151,7 +156,7 @@ int PMPI_Startall(int count, MPI_Request * requests) if (not TRACE_smpi_view_internals()) for (int i = 0; i < count; i++) { - req = requests[i]; + MPI_Request req = requests[i]; if (req->flags() & MPI_REQ_RECV) TRACE_smpi_recv(getPid(req->comm(), req->src()), my_proc_id, req->tag()); } @@ -167,10 +172,9 @@ int PMPI_Request_free(MPI_Request * request) int retval = 0; smpi_bench_end(); - if (*request == MPI_REQUEST_NULL) { - retval = MPI_ERR_ARG; - } else { + if (*request != MPI_REQUEST_NULL) { simgrid::smpi::Request::unref(request); + *request = MPI_REQUEST_NULL; retval = MPI_SUCCESS; } smpi_bench_begin(); @@ -262,6 +266,12 @@ int PMPI_Isend(const void *buf, int count, MPI_Datatype datatype, int dst, int t return retval; } +int PMPI_Irsend(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm, + MPI_Request* request) +{ + return PMPI_Isend(buf, count, datatype, dst, tag, comm, request); +} + int PMPI_Issend(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm, MPI_Request* request) { int retval = 0; @@ -390,6 +400,135 @@ int PMPI_Send(const void *buf, int count, MPI_Datatype datatype, int dst, int ta return retval; } +int PMPI_Rsend(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm) +{ + return PMPI_Send(buf, count, datatype, dst, tag, comm); +} + +int PMPI_Bsend(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm) +{ + int retval = 0; + + smpi_bench_end(); + + if (comm == MPI_COMM_NULL) { + retval = MPI_ERR_COMM; + } else if (dst == MPI_PROC_NULL) { + retval = MPI_SUCCESS; + } else if (dst >= comm->group()->size() || dst <0){ + retval = MPI_ERR_RANK; + } else if ((count < 0) || (buf == nullptr && count > 0)) { + retval = MPI_ERR_COUNT; + } else if (datatype==MPI_DATATYPE_NULL || not datatype->is_valid()) { + retval = MPI_ERR_TYPE; + } else if(tag < 0 && tag != MPI_ANY_TAG){ + retval = MPI_ERR_TAG; + } else { + int my_proc_id = simgrid::s4u::this_actor::get_pid(); + int dst_traced = getPid(comm, dst); + int bsend_buf_size = 0; + void* bsend_buf = nullptr; + smpi_process()->bsend_buffer(&bsend_buf, &bsend_buf_size); + int size = datatype->get_extent() * count; + if(bsend_buf==nullptr || bsend_buf_size < size + MPI_BSEND_OVERHEAD ) + return MPI_ERR_BUFFER; + TRACE_smpi_comm_in(my_proc_id, __func__, + new simgrid::instr::Pt2PtTIData("bsend", dst, + datatype->is_replayable() ? count : count * datatype->size(), + tag, simgrid::smpi::Datatype::encode(datatype))); + if (not TRACE_smpi_view_internals()) { + TRACE_smpi_send(my_proc_id, my_proc_id, dst_traced, tag, count * datatype->size()); + } + + simgrid::smpi::Request::bsend(buf, count, datatype, dst, tag, comm); + retval = MPI_SUCCESS; + + TRACE_smpi_comm_out(my_proc_id); + } + + smpi_bench_begin(); + return retval; +} + +int PMPI_Ibsend(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm, MPI_Request* request) +{ + int retval = 0; + + smpi_bench_end(); + if (request == nullptr) { + retval = MPI_ERR_ARG; + } else if (comm == MPI_COMM_NULL) { + retval = MPI_ERR_COMM; + } else if (dst == MPI_PROC_NULL) { + *request = MPI_REQUEST_NULL; + retval = MPI_SUCCESS; + } else if (dst >= comm->group()->size() || dst <0){ + retval = MPI_ERR_RANK; + } else if ((count < 0) || (buf==nullptr && count > 0)) { + retval = MPI_ERR_COUNT; + } else if (datatype==MPI_DATATYPE_NULL || not datatype->is_valid()) { + retval = MPI_ERR_TYPE; + } else if(tag<0 && tag != MPI_ANY_TAG){ + retval = MPI_ERR_TAG; + } else { + int my_proc_id = simgrid::s4u::this_actor::get_pid(); + int trace_dst = getPid(comm, dst); + int bsend_buf_size = 0; + void* bsend_buf = nullptr; + smpi_process()->bsend_buffer(&bsend_buf, &bsend_buf_size); + int size = datatype->get_extent() * count; + if(bsend_buf==nullptr || bsend_buf_size < size + MPI_BSEND_OVERHEAD ) + return MPI_ERR_BUFFER; + TRACE_smpi_comm_in(my_proc_id, __func__, + new simgrid::instr::Pt2PtTIData("ibsend", dst, + datatype->is_replayable() ? count : count * datatype->size(), + tag, simgrid::smpi::Datatype::encode(datatype))); + + TRACE_smpi_send(my_proc_id, my_proc_id, trace_dst, tag, count * datatype->size()); + + *request = simgrid::smpi::Request::ibsend(buf, count, datatype, dst, tag, comm); + retval = MPI_SUCCESS; + + TRACE_smpi_comm_out(my_proc_id); + } + + smpi_bench_begin(); + if (retval != MPI_SUCCESS && request!=nullptr) + *request = MPI_REQUEST_NULL; + return retval; +} + +int PMPI_Bsend_init(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm, MPI_Request* request) +{ + + int retval = 0; + + smpi_bench_end(); + if (request == nullptr) { + retval = MPI_ERR_ARG; + } else if (comm == MPI_COMM_NULL) { + retval = MPI_ERR_COMM; + } else if (datatype==MPI_DATATYPE_NULL || not datatype->is_valid()) { + retval = MPI_ERR_TYPE; + } else if (dst == MPI_PROC_NULL) { + retval = MPI_SUCCESS; + } else { + int bsend_buf_size = 0; + void* bsend_buf = nullptr; + smpi_process()->bsend_buffer(&bsend_buf, &bsend_buf_size); + if( bsend_buf==nullptr || bsend_buf_size < datatype->get_extent() * count + MPI_BSEND_OVERHEAD ) { + retval = MPI_ERR_BUFFER; + } else { + *request = simgrid::smpi::Request::bsend_init(buf, count, datatype, dst, tag, comm); + retval = MPI_SUCCESS; + } + } + smpi_bench_begin(); + if (retval != MPI_SUCCESS && request != nullptr) + *request = MPI_REQUEST_NULL; + return retval; +} + int PMPI_Ssend(const void* buf, int count, MPI_Datatype datatype, int dst, int tag, MPI_Comm comm) { int retval = 0; @@ -656,7 +795,7 @@ int PMPI_Wait(MPI_Request * request, MPI_Status * status) } else if (*request == MPI_REQUEST_NULL) { retval = MPI_SUCCESS; } else { - //for tracing, save the handle which might get overriden before we can use the helper on it + // for tracing, save the handle which might get overridden before we can use the helper on it MPI_Request savedreq = *request; if (savedreq != MPI_REQUEST_NULL && not(savedreq->flags() & MPI_REQ_FINISHED) && not(savedreq->flags() & MPI_REQ_GENERALIZED)) @@ -692,7 +831,7 @@ int PMPI_Waitany(int count, MPI_Request requests[], int *index, MPI_Status * sta return MPI_SUCCESS; smpi_bench_end(); - //for tracing, save the handles which might get overriden before we can use the helper on it + // for tracing, save the handles which might get overridden before we can use the helper on it std::vector savedreqs(requests, requests + count); for (MPI_Request& req : savedreqs) { if (req != MPI_REQUEST_NULL && not(req->flags() & MPI_REQ_FINISHED)) @@ -723,7 +862,7 @@ int PMPI_Waitall(int count, MPI_Request requests[], MPI_Status status[]) { smpi_bench_end(); - //for tracing, save the handles which might get overriden before we can use the helper on it + // for tracing, save the handles which might get overridden before we can use the helper on it std::vector savedreqs(requests, requests + count); for (MPI_Request& req : savedreqs) { if (req != MPI_REQUEST_NULL && not(req->flags() & MPI_REQ_FINISHED)) @@ -813,12 +952,11 @@ int PMPI_Grequest_complete( MPI_Request request){ return simgrid::smpi::Request::grequest_complete(request); } - int PMPI_Request_get_status( MPI_Request request, int *flag, MPI_Status *status){ if(request==MPI_REQUEST_NULL){ *flag=1; simgrid::smpi::Status::empty(status); - return MPI_ERR_REQUEST; + return MPI_SUCCESS; } else if (flag==NULL || status ==NULL){ return MPI_ERR_ARG; } @@ -826,9 +964,13 @@ int PMPI_Request_get_status( MPI_Request request, int *flag, MPI_Status *status) } MPI_Request PMPI_Request_f2c(MPI_Fint request){ + if(request==-1) + return MPI_REQUEST_NULL; return static_cast(simgrid::smpi::Request::f2c(request)); } MPI_Fint PMPI_Request_c2f(MPI_Request request) { + if(request==MPI_REQUEST_NULL) + return -1; return request->c2f(); }