X-Git-Url: http://bilbo.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/b7ed19dfcc221d7b3eca182abb5c4a3946671172..8be89720f55b4ceeb2877531ae1602cc7ed947d6:/src/smpi/colls/bcast/bcast-SMP-linear.cpp diff --git a/src/smpi/colls/bcast/bcast-SMP-linear.cpp b/src/smpi/colls/bcast/bcast-SMP-linear.cpp index 2b3aee4023..c35df9916a 100644 --- a/src/smpi/colls/bcast/bcast-SMP-linear.cpp +++ b/src/smpi/colls/bcast/bcast-SMP-linear.cpp @@ -1,22 +1,20 @@ -/* Copyright (c) 2013-2014. The SimGrid Team. +/* Copyright (c) 2013-2023. The SimGrid Team. * All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ -#include "../colls_private.h" +#include "../colls_private.hpp" int bcast_SMP_linear_segment_byte = 8192; - -int Coll_bcast_SMP_linear::bcast(void *buf, int count, - MPI_Datatype datatype, int root, - MPI_Comm comm) +namespace simgrid::smpi { +int bcast__SMP_linear(void *buf, int count, + MPI_Datatype datatype, int root, + MPI_Comm comm) { int tag = COLL_TAG_BCAST; MPI_Status status; MPI_Request request; - MPI_Request *request_array; - MPI_Status *status_array; int rank, size; int i; MPI_Aint extent; @@ -32,12 +30,11 @@ int Coll_bcast_SMP_linear::bcast(void *buf, int count, num_core = comm->get_intra_comm()->size(); }else{ //implementation buggy in this case - return Coll_bcast_mpich::bcast( buf , count, datatype, - root, comm); + return bcast__mpich(buf, count, datatype, root, comm); } int segment = bcast_SMP_linear_segment_byte / extent; - segment = segment == 0 ? 1 :segment; + segment = segment == 0 ? 1 :segment; int pipe_length = count / segment; int remainder = count % segment; int increment = segment * extent; @@ -52,9 +49,9 @@ int Coll_bcast_SMP_linear::bcast(void *buf, int count, // call native when MPI communication size is too small if (size <= num_core) { - XBT_WARN("MPI_bcast_SMP_linear use default MPI_bcast."); - Coll_bcast_default::bcast(buf, count, datatype, root, comm); - return MPI_SUCCESS; + XBT_INFO("size <= num_core : MPI_bcast_SMP_linear use default MPI_bcast."); + bcast__default(buf, count, datatype, root, comm); + return MPI_SUCCESS; } // if root is not zero send to rank zero first if (root != 0) { @@ -63,7 +60,7 @@ int Coll_bcast_SMP_linear::bcast(void *buf, int count, else if (rank == 0) Request::recv(buf, count, datatype, root, tag, comm, &status); } - // when a message is smaller than a block size => no pipeline + // when a message is smaller than a block size => no pipeline if (count <= segment) { // case ROOT if (rank == 0) { @@ -98,10 +95,8 @@ int Coll_bcast_SMP_linear::bcast(void *buf, int count, } // pipeline bcast else { - request_array = - (MPI_Request *) xbt_malloc((size + pipe_length) * sizeof(MPI_Request)); - status_array = - (MPI_Status *) xbt_malloc((size + pipe_length) * sizeof(MPI_Status)); + auto* request_array = new MPI_Request[size + pipe_length]; + auto* status_array = new MPI_Status[size + pipe_length]; // case ROOT of each SMP if (rank % num_core == 0) { @@ -163,16 +158,16 @@ int Coll_bcast_SMP_linear::bcast(void *buf, int count, } } } - free(request_array); - free(status_array); + delete[] request_array; + delete[] status_array; } - // when count is not divisible by block size, use default BCAST for the remainder if ((remainder != 0) && (count > segment)) { - XBT_WARN("MPI_bcast_SMP_linear use default MPI_bcast."); - Colls::bcast((char *) buf + (pipe_length * increment), remainder, datatype, - root, comm); + XBT_INFO("MPI_bcast_SMP_linear: count is not divisible by block size, use default MPI_bcast for remainder."); + colls::bcast((char*)buf + (pipe_length * increment), remainder, datatype, root, comm); } return MPI_SUCCESS; } + +} // namespace simgrid::smpi