-/* Copyright (c) 2009-2019. The SimGrid Team. All rights reserved. */
+/* Copyright (c) 2009-2021. The SimGrid Team. All rights reserved. */
/* This program is free software; you can redistribute it and/or modify it
* under the terms of the license (GNU LGPL) which comes with this package. */
#include "src/smpi/include/smpi_actor.hpp"
#include "mc/mc.h"
+#include "simgrid/s4u/Engine.hpp"
+#include "simgrid/s4u/Mutex.hpp"
#include "smpi_comm.hpp"
+#include "smpi_info.hpp"
#include "src/mc/mc_replay.hpp"
#include "src/simix/smx_private.hpp"
#if HAVE_PAPI
#include "papi.h"
-extern std::string papi_default_config_name;
#endif
XBT_LOG_NEW_DEFAULT_SUBCATEGORY(smpi_process, smpi, "Logging specific to SMPI (kernel)");
namespace simgrid {
namespace smpi {
+simgrid::xbt::Extension<simgrid::s4u::Actor, ActorExt> ActorExt::EXTENSION_ID;
-ActorExt::ActorExt(s4u::ActorPtr actor) : actor_(actor)
+ActorExt::ActorExt(s4u::Actor* actor) : actor_(actor)
{
+ if (not simgrid::smpi::ActorExt::EXTENSION_ID.valid())
+ simgrid::smpi::ActorExt::EXTENSION_ID = simgrid::s4u::Actor::extension_create<simgrid::smpi::ActorExt>();
+
mailbox_ = s4u::Mailbox::by_name("SMPI-" + std::to_string(actor_->get_pid()));
mailbox_small_ = s4u::Mailbox::by_name("small-" + std::to_string(actor_->get_pid()));
mailboxes_mutex_ = s4u::Mutex::create();
MC_ignore_heap(timer_, xbt_os_timer_size());
#if HAVE_PAPI
- if (not simgrid::config::get_value<std::string>("smpi/papi-events").empty()) {
+ if (not smpi_cfg_papi_events_file().empty()) {
// TODO: Implement host/process/thread based counters. This implementation
// just always takes the values passed via "default", like this:
// "default:COUNTER1:COUNTER2:COUNTER3;".
- auto it = units2papi_setup.find(papi_default_config_name);
+ auto it = units2papi_setup.find("default");
if (it != units2papi_setup.end()) {
papi_event_set_ = it->second.event_set;
papi_counter_data_ = it->second.counter_data;
ActorExt::~ActorExt()
{
- TRACE_smpi_finalize(actor_->get_pid());
-
- if (comm_self_ != MPI_COMM_NULL)
- simgrid::smpi::Comm::destroy(comm_self_);
- if (comm_intra_ != MPI_COMM_NULL)
- simgrid::smpi::Comm::destroy(comm_intra_);
xbt_os_timer_free(timer_);
}
{
state_ = SmpiProcessState::FINALIZED;
XBT_DEBUG("<%ld> Process left the game", actor_->get_pid());
-
+ if (info_env_ != MPI_INFO_NULL)
+ simgrid::smpi::Info::unref(info_env_);
+ if (comm_self_ != MPI_COMM_NULL)
+ simgrid::smpi::Comm::destroy(comm_self_);
+ if (comm_intra_ != MPI_COMM_NULL)
+ simgrid::smpi::Comm::destroy(comm_intra_);
smpi_deployment_unregister_process(instance_id_);
}
/** @brief Check if a process is finalized */
-int ActorExt::finalized()
+int ActorExt::finalized() const
{
return (state_ == SmpiProcessState::FINALIZED);
}
/** @brief Check if a process is partially initialized already */
-int ActorExt::initializing()
+int ActorExt::initializing() const
{
return (state_ == SmpiProcessState::INITIALIZING);
}
/** @brief Check if a process is initialized */
-int ActorExt::initialized()
+int ActorExt::initialized() const
{
// TODO cheinrich: Check if we still need this. This should be a global condition, not for a
// single process ... ?
replaying_ = value;
}
-bool ActorExt::replaying()
+bool ActorExt::replaying() const
{
return replaying_;
}
privatized_region_ = region;
}
-smpi_privatization_region_t ActorExt::privatized_region()
+smpi_privatization_region_t ActorExt::privatized_region() const
{
return privatized_region_;
}
-MPI_Comm ActorExt::comm_world()
+MPI_Comm ActorExt::comm_world() const
{
return comm_world_ == nullptr ? MPI_COMM_NULL : *comm_world_;
}
-s4u::MutexPtr ActorExt::mailboxes_mutex()
+s4u::MutexPtr ActorExt::mailboxes_mutex() const
{
return mailboxes_mutex_;
}
#if HAVE_PAPI
-int ActorExt::papi_event_set()
+int ActorExt::papi_event_set() const
{
return papi_event_set_;
}
void ActorExt::simulated_start()
{
- simulated_ = SIMIX_get_clock();
+ simulated_ = s4u::Engine::get_clock();
}
-double ActorExt::simulated_elapsed()
+double ActorExt::simulated_elapsed() const
{
- return SIMIX_get_clock() - simulated_;
+ return s4u::Engine::get_clock() - simulated_;
}
MPI_Comm ActorExt::comm_self()
{
if (comm_self_ == MPI_COMM_NULL) {
- MPI_Group group = new Group(1);
- comm_self_ = new Comm(group, nullptr);
- group->set_mapping(actor_, 0);
+ auto* group = new Group(1);
+ comm_self_ = new Comm(group, nullptr);
+ comm_self_->set_name("MPI_COMM_SELF");
+ group->set_mapping(actor_->get_pid(), 0);
}
return comm_self_;
}
MPI_Info ActorExt::info_env()
{
+ if (info_env_==MPI_INFO_NULL)
+ info_env_=new Info();
return info_env_;
}
sampling_ = s;
}
-int ActorExt::sampling()
+int ActorExt::sampling() const
{
return sampling_;
}
xbt_assert(smpi_get_universe_size() != 0, "SimGrid was not initialized properly before entering MPI_Init. "
"Aborting, please check compilation process and use smpirun.");
- simgrid::s4u::ActorPtr self = simgrid::s4u::Actor::self();
- // cheinrich: I'm not sure what the impact of the SMPI_switch_data_segment on this call is. I moved
- // this up here so that I can set the privatized region before the switch.
ActorExt* ext = smpi_process();
// if we are in MPI_Init and argc handling has already been done.
if (ext->initialized())
return;
- if (smpi_privatize_global_variables == SmpiPrivStrategies::MMAP) {
- /* Now using the segment index of this process */
- ext->set_privatized_region(smpi_init_global_memory_segment_process());
- /* Done at the process's creation */
- SMPI_switch_data_segment(self);
- }
-
+ const simgrid::s4u::Actor* self = simgrid::s4u::Actor::self();
ext->instance_id_ = self->get_property("instance_id");
const int rank = xbt_str_parse_int(self->get_property("rank"), "Cannot parse rank");
// set the process attached to the mailbox
ext->mailbox_small_->set_receiver(ext->actor_);
- XBT_DEBUG("<%ld> SMPI process has been initialized: %p", ext->actor_->get_pid(), ext->actor_.get());
+ XBT_DEBUG("<%ld> SMPI process has been initialized: %p", ext->actor_->get_pid(), ext->actor_);
}
-int ActorExt::get_optind()
+int ActorExt::get_optind() const
{
return optind_;
}
optind_ = new_optind;
}
+void ActorExt::bsend_buffer(void** buf, int* size)
+{
+ *buf = bsend_buffer_;
+ *size = bsend_buffer_size_;
+}
+
+int ActorExt::set_bsend_buffer(void* buf, int size)
+{
+ if(buf!=nullptr && bsend_buffer_!=nullptr)
+ return MPI_ERR_BUFFER;
+ bsend_buffer_ = buf;
+ bsend_buffer_size_= size;
+ return MPI_SUCCESS;
+}
+
} // namespace smpi
} // namespace simgrid