]> AND Private Git Repository - loba.git/blobdiff - process.cpp
Logo AND Algorithmique Numérique Distribuée

Private GIT Repository
Update TODO list.
[loba.git] / process.cpp
index 34ddbea2b713d3071ee24d83bda4ef7c3269c6e9..33a993ef5eb8e96041302b563acef1fafeb557b7 100644 (file)
@@ -1,6 +1,7 @@
 #include <algorithm>
 #include <tr1/functional>
 #include <iterator>
 #include <algorithm>
 #include <tr1/functional>
 #include <iterator>
+#include <numeric>
 #include <stdexcept>
 #include <sstream>
 #include <xbt/log.h>
 #include <stdexcept>
 #include <sstream>
 #include <xbt/log.h>
@@ -18,6 +19,18 @@ double process::total_load_init = 0.0;
 double process::total_load_running = 0.0;
 double process::total_load_exit = 0.0;
 
 double process::total_load_running = 0.0;
 double process::total_load_exit = 0.0;
 
+namespace {
+
+    void sleep_until_date(double& date, double duration = 0.0)
+    {
+        double sleep_duration = date - MSG_get_clock();
+        if (sleep_duration > 0.0)
+            MSG_process_sleep(sleep_duration);
+        date = MSG_get_clock() + duration;
+    }
+
+}
+
 process::process(int argc, char* argv[])
 {
     if (argc < 2 || !(std::istringstream(argv[1]) >> real_load))
 process::process(int argc, char* argv[])
 {
     if (argc < 2 || !(std::istringstream(argv[1]) >> real_load))
@@ -46,6 +59,10 @@ process::process(int argc, char* argv[])
 
     comp_iter = lb_iter = 0;
 
 
     comp_iter = lb_iter = 0;
 
+    lb_thread = new_msg_thread("loba",
+                               std::tr1::bind(&process::load_balance_loop,
+                                              this));
+
     e_xbt_log_priority_t logp = xbt_log_priority_verbose;
     if (!LOG_ISENABLED(logp))
         return;
     e_xbt_log_priority_t logp = xbt_log_priority_verbose;
     if (!LOG_ISENABLED(logp))
         return;
@@ -64,124 +81,115 @@ process::process(int argc, char* argv[])
 
 process::~process()
 {
 
 process::~process()
 {
+    delete lb_thread;
     total_load_exit += real_load;
     total_load_exit += real_load;
-    if (opt::bookkeeping) {
-        XBT_INFO("Final load after %d:%d iterations: %g ; expected: %g",
-                 lb_iter, comp_iter, real_load, expected_load);
-    } else {
-        XBT_INFO("Final load after %d iterations: %g",
-                 lb_iter, real_load);
-        if (lb_iter != comp_iter)
-            XBT_WARN("lb_iter (%d) and comp_iter (%d) differ!",
-                     lb_iter, comp_iter);
-    }
+    if (opt::log_rate < 0)
+        return;
+    XBT_INFO("Final load after %d:%d iterations: %g",
+             lb_iter, comp_iter, real_load);
+    XBT_VERB("Expected load was: %g", expected_load);
     XBT_VERB("Total computation for this process: %g", comp);
 }
 
 int process::run()
 {
     XBT_VERB("Total computation for this process: %g", comp);
 }
 
 int process::run()
 {
-    double next_iter_after_date = 0.0;
-    XBT_INFO("Initial load: %g", real_load);
+    if (opt::log_rate >= 0) {
+        XBT_INFO("Initial load: %g", real_load);
+        XBT_VERB("Initial expected load: %g", expected_load);
+    }
     XBT_VERB("Starting...");
     XBT_VERB("Starting...");
-    while (true) {
-        if (get_load() > 0.0) {
-            double now = MSG_get_clock();
-            if (now < next_iter_after_date)
-                MSG_process_sleep(next_iter_after_date - now);
-            next_iter_after_date = MSG_get_clock() + opt::min_lb_iter_duration;
-
-            ++lb_iter;
-
-            if (opt::log_rate && lb_iter % opt::log_rate == 0) {
-                if (opt::bookkeeping)
-                    XBT_INFO("(%u:%u) current load: %g ; expected: %g",
-                             lb_iter, comp_iter, real_load, expected_load);
-                else
-                    XBT_INFO("(%u) current load: %g",
-                             lb_iter, real_load);
-            }
+    mutex.acquire();
+    lb_thread->start();
+    while (lb_iter <= opt::comp_iter_delay)
+        cond.wait(mutex);
+    mutex.release();
+    double sleep_duration = opt::comp_time_delay - MSG_get_clock();
+    if (sleep_duration > 0.0)
+        MSG_process_sleep(sleep_duration);
+    compute_loop();
+    lb_thread->wait();
+    XBT_VERB("Done.");
+    return 0;
+}
 
 
-            load_balance();
+void process::load_balance_loop()
+{
+    using std::tr1::bind;
+    using std::tr1::placeholders::_1;
 
 
-            print_loads(true, xbt_log_priority_debug);
+    double next_iter_after_date = MSG_get_clock() + opt::min_lb_iter_duration;
+    while (still_running()) {
+        if (lb_iter == opt::comp_iter_delay) {
+            mutex.acquire();
+            ++lb_iter;
+            cond.signal();
+            mutex.release();
+        } else {
+            ++lb_iter;
         }
 
         }
 
-        // send load information, and load (data) if any
-        send_all();
-        if (real_load > 0.0) {
-            ++comp_iter;
-            compute();
-        }
+        if (!opt::bookkeeping)
+            expected_load = real_load - get_sum_of_to_send();
+        // nothing to do with opt::bookkeeping
 
 
-        if (opt::lb_maxiter && lb_iter >= opt::lb_maxiter) {
-            XBT_VERB("Reached lb_maxiter: %d/%d", lb_iter, opt::lb_maxiter);
-            break;
-        }
-        if (opt::comp_maxiter && comp_iter >= opt::comp_maxiter) {
-            XBT_VERB("Reached comp_maxiter: %d/%d", comp_iter, opt::comp_maxiter);
-            break;
-        }
-        if (opt::time_limit && MSG_get_clock() >= opt::time_limit) {
-            XBT_VERB("Reached time limit: %g/%g", MSG_get_clock(), opt::time_limit);
-            break;
+        if (opt::log_rate && lb_iter % opt::log_rate == 0) {
+            XBT_INFO("(%u:%u) current load: %g", lb_iter, comp_iter, real_load);
+            XBT_VERB("... expected load: %g", expected_load);
         }
 
         }
 
-        // block on receiving unless there is something to compute or
-        // to send
-        double timeout;
-        if (real_load != 0 || get_load() != prev_load_broadcast)
-            timeout = 0.0;
-        else if (opt::min_lb_iter_duration)
-            timeout = opt::min_lb_iter_duration;
-        else
-            timeout = 1.0;
-        receive(timeout);
+        if (expected_load > 0.0)
+            load_balance();
 
 
-        // one of our neighbor is finalizing
-        if (opt::exit_on_close && close_received) {
-            XBT_VERB("Close received");
-            break;
-        }
+        print_loads(true, xbt_log_priority_debug);
 
 
-        // have no load and cannot receive anything
-        if (real_load == 0.0 && !may_receive()) {
-            XBT_VERB("I'm a poor lonesome process, and I have no load...");
-            break;
-        }
+        // send
+        std::for_each(neigh.begin(), neigh.end(),
+                      bind(&process::ctrl_send, this, _1));
+        prev_load_broadcast = expected_load;
 
 
-        // fixme: this check should be implemented with a distributed
-        // algorithm, and not a shared global variable!
-        // fixme: should this chunk be moved before call to receive() ?
-        if (100.0 * total_load_running / total_load_init <=
-            opt::load_ratio_threshold) {
-            XBT_VERB("No more load to balance in system.");
-            break;
-        } else {
-            XBT_DEBUG("still %g load to balance, continuing...", total_load_running);
-        }
-    }
-    XBT_VERB("Going to finalize...");
-    finalize();
+        sleep_until_date(next_iter_after_date, opt::min_lb_iter_duration);
+        ctrl_receive(0.0);
 
 
-    /* Open Questions :
-     * - definition of load on heterogeneous hosts ?
-     * - how to detect convergence ?
-     * - how to manage link failures ?
-     */
+        comm.ctrl_flush(false);
+    }
 
 
-    XBT_VERB("Done.");
-    return 0;
+    XBT_VERB("Going to finalize for %s...", __func__);
+    XBT_DEBUG("send CTRL_CLOSE to %zu neighbor%s",
+              neigh.size(), ESSE(neigh.size()));
+    std::for_each(neigh.begin(), neigh.end(),
+                  bind(&process::ctrl_close, this, _1));
+    while (ctrl_close_pending) {
+        comm.ctrl_flush(false);
+        XBT_DEBUG("waiting for %d CTRL CLOSE", ctrl_close_pending);
+        ctrl_receive(-1.0);
+    }
+    comm.ctrl_flush(true);
 }
 
 }
 
-void process::load_balance()
+void process::compute_loop()
 {
 {
-    if (lb_iter == 1)           // warn only once
-        XBT_WARN("process::load_balance() is a no-op!");
-}
+    using std::tr1::bind;
+    using std::tr1::placeholders::_1;
 
 
-void process::compute()
-{
-    if (real_load > 0.0) {
+    double next_iter_after_date = MSG_get_clock() + opt::min_comp_iter_duration;
+    while (still_running()) {
+        // receive
+        if (real_load > 0.0)
+            data_receive(0.0);
+        else
+            data_receive(opt::min_comp_iter_duration);
+
+        comm.data_flush(false);
+
+        if (real_load == 0.0)
+            continue;
+
+        // send
+        std::for_each(neigh.begin(), neigh.end(),
+                      bind(&process::data_send, this, _1));
+
+        // compute
+        ++comp_iter;
         double flops = opt::comp_cost(real_load);
         m_task_t task = MSG_task_create("computation", flops, 0.0, NULL);
         TRACE_msg_set_task_category(task, TRACE_CAT_COMP);
         double flops = opt::comp_cost(real_load);
         m_task_t task = MSG_task_create("computation", flops, 0.0, NULL);
         TRACE_msg_set_task_category(task, TRACE_CAT_COMP);
@@ -189,146 +197,208 @@ void process::compute()
         MSG_task_execute(task);
         comp += flops;
         MSG_task_destroy(task);
         MSG_task_execute(task);
         comp += flops;
         MSG_task_destroy(task);
-    } else {
-        XBT_DEBUG("nothing to compute !");
+
+        sleep_until_date(next_iter_after_date, opt::min_comp_iter_duration);
+    }
+
+    XBT_VERB("Going to finalize for %s...", __func__);
+    // last send, for not losing load scheduled to be sent
+    std::for_each(neigh.begin(), neigh.end(),
+                  bind(&process::data_send, this, _1));
+    finalizing = true;
+    total_load_running -= real_load;
+    XBT_DEBUG("send DATA_CLOSE to %zu neighbor%s",
+              neigh.size(), ESSE(neigh.size()));
+    std::for_each(neigh.begin(), neigh.end(),
+                  bind(&process::data_close, this, _1));
+    while (data_close_pending) {
+        comm.data_flush(false);
+        XBT_DEBUG("waiting for %d DATA CLOSE", data_close_pending);
+        data_receive(-1.0);
     }
     }
+    comm.data_flush(true);
+}
+
+bool process::still_running()
+{
+    static bool last_status = true;
+
+    if (!last_status) {
+        /* nop */
+
+    } else if (opt::time_limit && MSG_get_clock() >= opt::time_limit) {
+        XBT_VERB("Reached time limit: %g/%g", MSG_get_clock(), opt::time_limit);
+        last_status = false;
+
+    } else if (opt::lb_maxiter && lb_iter >= opt::lb_maxiter) {
+        XBT_VERB("Reached lb_maxiter: %d/%d", lb_iter, opt::lb_maxiter);
+        last_status = false;
+
+    } else if (opt::comp_maxiter && comp_iter >= opt::comp_maxiter) {
+        XBT_VERB("Reached comp_maxiter: %d/%d", comp_iter, opt::comp_maxiter);
+        last_status = false;
+
+    } else if (opt::exit_on_close && close_received) {
+        XBT_VERB("Close received");
+        last_status = false;
+
+    } else if (real_load == 0.0 && !data_close_pending) {
+        XBT_VERB("I'm a poor lonesome process, and I have no load...");
+        last_status = false;
+
+    } else if (100.0 * total_load_running / total_load_init <=
+               opt::load_ratio_threshold) {
+        // fixme: this check should be implemented with a distributed
+        // algorithm, and not a shared global variable!
+        XBT_VERB("No more load to balance in system.");
+        last_status = false;
+    }
+
+    return last_status;
+}
+
+double process::get_sum_of_to_send() const
+{
+    using std::tr1::bind;
+    using std::tr1::placeholders::_1;
+    using std::tr1::placeholders::_2;
+
+    return std::accumulate(neigh.begin(), neigh.end(), 0.0,
+                           bind(std::plus<double>(),
+                                _1, bind(&neighbor::get_to_send, _2)));
+}
+
+void process::load_balance()
+{
+    if (lb_iter == 1)           // warn only once
+        XBT_WARN("process::load_balance() is a no-op!");
 }
 
 void process::send(neighbor& nb, double amount)
 {
 }
 
 void process::send(neighbor& nb, double amount)
 {
-    set_load(get_load() - amount);
+    expected_load -= amount;
     nb.set_to_send(nb.get_to_send() + amount);
     nb.set_to_send(nb.get_to_send() + amount);
-    nb.set_load(nb.get_load() + amount); // fixme: make this optional?
+    nb.set_load(nb.get_load() + amount);
 }
 
 }
 
-void process::send1_no_bookkeeping(neighbor& nb)
+void process::ctrl_send(neighbor& nb)
 {
 {
-    if (real_load != prev_load_broadcast)
-        comm.send(nb.get_ctrl_mbox(), new message(message::INFO, real_load));
-    double load_to_send = nb.get_to_send();
-    if (load_to_send > 0.0) {
-        comm.send(nb.get_data_mbox(), new message(message::LOAD, load_to_send));
-        nb.set_to_send(0.0);
+    double info_to_send = expected_load;
+    if (info_to_send != prev_load_broadcast)
+        comm.ctrl_send(nb.get_ctrl_mbox(),
+                       new message(message::INFO, info_to_send));
+    if (opt::bookkeeping) {
+        double debt_to_send = nb.get_to_send();
+        if (debt_to_send > 0.0) {
+            nb.set_to_send(0.0);
+            nb.set_debt(nb.get_debt() + debt_to_send);
+            comm.ctrl_send(nb.get_ctrl_mbox(),
+                           new message(message::CREDIT, debt_to_send));
+        }
     }
 }
 
     }
 }
 
-void process::send1_bookkeeping(neighbor& nb)
+void process::data_send(neighbor& nb)
 {
 {
-    if (expected_load != prev_load_broadcast)
-        comm.send(nb.get_ctrl_mbox(),
-                  new message(message::INFO, expected_load));
     double load_to_send;
     double load_to_send;
-    double new_debt;
-    double debt_to_send = nb.get_to_send();
-    if (debt_to_send > 0.0) {
-        comm.send(nb.get_ctrl_mbox(),
-                  new message(message::CREDIT, debt_to_send));
-        nb.set_to_send(0.0);
-        new_debt = nb.get_debt() + debt_to_send;
+    if (opt::bookkeeping) {
+        load_to_send = std::min(real_load, nb.get_debt());
+        if (load_to_send >= opt::min_transfer_amount) {
+            nb.set_debt(nb.get_debt() - load_to_send);
+            real_load -= load_to_send;
+        } else {
+            load_to_send = 0.0;
+        }
     } else {
     } else {
-        new_debt = nb.get_debt();
+        load_to_send = nb.get_to_send();
+        if (load_to_send >= opt::min_transfer_amount) {
+            nb.set_to_send(0.0);
+            real_load -= load_to_send;
+        } else {
+            load_to_send = 0.0;
+        }
     }
     }
-    if (real_load <= new_debt) {
-        load_to_send = real_load;
-        nb.set_debt(new_debt - load_to_send);
-        real_load = 0.0;
-    } else {
-        load_to_send = new_debt;
-        nb.set_debt(0.0);
-        real_load -= load_to_send;
+    while (load_to_send > 0.0) {
+        double amount;
+        if (opt::max_transfer_amount)
+            amount = std::min(load_to_send, opt::max_transfer_amount);
+        else
+            amount = load_to_send;
+        comm.data_send(nb.get_data_mbox(), new message(message::LOAD, amount));
+        load_to_send -= amount;
     }
     }
-    if (load_to_send > 0.0)
-        comm.send(nb.get_data_mbox(), new message(message::LOAD, load_to_send));
 }
 
 }
 
-void process::send_all()
+void process::ctrl_close(neighbor& nb)
 {
 {
-    using std::tr1::bind;
-    using std::tr1::placeholders::_1;
+    comm.ctrl_send(nb.get_ctrl_mbox(), new message(message::CTRL_CLOSE, 0.0));
+}
 
 
-    if (opt::bookkeeping) {
-        std::for_each(neigh.begin(), neigh.end(),
-                      bind(&process::send1_bookkeeping, this, _1));
-        prev_load_broadcast = expected_load;
-    } else {
-        std::for_each(neigh.begin(), neigh.end(),
-                      bind(&process::send1_no_bookkeeping, this, _1));
-        prev_load_broadcast = real_load;
-    }
-    comm.flush(false);
+void process::data_close(neighbor& nb)
+{
+    comm.data_send(nb.get_data_mbox(), new message(message::DATA_CLOSE, 0.0));
 }
 
 }
 
-void process::receive(double timeout)
+void process::ctrl_receive(double timeout)
 {
     message* msg;
     m_host_t from;
 
 {
     message* msg;
     m_host_t from;
 
-    XBT_DEBUG("%sblocking receive (%g)", "\0non-" + !timeout, timeout);
-    while (may_receive() && comm.recv(msg, from, timeout)) {
-        switch (msg->get_type()) {
-        case message::INFO: {
-            neighbor* n = rev_neigh[from];
-            n->set_load(msg->get_amount());
-            break;
-        }
-        case message::CREDIT:
-            expected_load += msg->get_amount();
-            break;
-        case message::LOAD: {
-            double ld = msg->get_amount();
-            real_load += ld;
-            if (finalizing)
-                total_load_running -= ld;
-            break;
-        }
-        case message::CTRL_CLOSE:
-            ctrl_close_pending--;
-            close_received = true;
-            break;
-        case message::DATA_CLOSE:
-            data_close_pending--;
-            close_received = true;
-            break;
-        }
-        delete msg;
-        timeout = 0.0;          // only wait on first recv
+    XBT_DEBUG("%sblocking receive on ctrl (%g)", "\0non-" + !timeout, timeout);
+    while (ctrl_close_pending && comm.ctrl_recv(msg, from, timeout)) {
+        handle_message(msg, from);
+        timeout = 0.0;
     }
     }
-    comm.flush(false);
 }
 
 }
 
-void process::finalize1(neighbor& nb)
+void process::data_receive(double timeout)
 {
 {
-    comm.send(nb.get_ctrl_mbox(), new message(message::CTRL_CLOSE, 0.0));
-    comm.send(nb.get_data_mbox(), new message(message::DATA_CLOSE, 0.0));
+    message* msg;
+    m_host_t from;
+
+    XBT_DEBUG("%sblocking receive on data (%g)", "\0non-" + !timeout, timeout);
+    while (data_close_pending && comm.data_recv(msg, from, timeout)) {
+        handle_message(msg, from);
+        timeout = 0.0;
+    }
 }
 
 }
 
-void process::finalize()
+void process::handle_message(message* msg, m_host_t from)
 {
 {
-    using std::tr1::bind;
-    using std::tr1::placeholders::_1;
-
-    finalizing = true;
-    total_load_running -= real_load;
-
-    XBT_DEBUG("send CLOSE to %zu neighbor%s", neigh.size(), ESSE(neigh.size()));
-    std::for_each(neigh.begin(), neigh.end(),
-                  bind(&process::finalize1, this, _1));
-
-    while (may_receive()) {
-        comm.flush(false);
-        XBT_DEBUG("waiting for %d CTRL and %d DATA CLOSE",
-                  ctrl_close_pending, data_close_pending);
-        receive(-1.0);
+    switch (msg->get_type()) {
+    case message::INFO: {
+        neighbor* n = rev_neigh[from];
+        n->set_load(msg->get_amount() + n->get_to_send());
+        break;
     }
     }
-
-    comm.flush(true);
+    case message::CREDIT:
+        expected_load += msg->get_amount();
+        break;
+    case message::LOAD: {
+        double ld = msg->get_amount();
+        real_load += ld;
+        if (finalizing)
+            total_load_running -= ld;
+        break;
+    }
+    case message::CTRL_CLOSE:
+        ctrl_close_pending--;
+        close_received = true;
+        break;
+    case message::DATA_CLOSE:
+        data_close_pending--;
+        close_received = true;
+        break;
+    }
+    delete msg;
 }
 
 #define print_loads_generic(vec, verbose, logp, cat)                    \
     if (_XBT_LOG_ISENABLEDV((*cat), logp)) {                            \
         using std::tr1::bind;                                           \
         using std::tr1::placeholders::_1;                               \
 }
 
 #define print_loads_generic(vec, verbose, logp, cat)                    \
     if (_XBT_LOG_ISENABLEDV((*cat), logp)) {                            \
         using std::tr1::bind;                                           \
         using std::tr1::placeholders::_1;                               \
-        XBT_XCLOG(cat, logp, "Neighbor loads:");                        \
+        XBT_XCLOG(cat, logp, "My load: %g (real); %g (expected).  "     \
+                  "Neighbor loads:", real_load, expected_load);         \
         std::for_each(vec.begin(), vec.end(),                           \
                       bind(&neighbor::print, _1, verbose, logp, cat));  \
     } else ((void)0)
         std::for_each(vec.begin(), vec.end(),                           \
                       bind(&neighbor::print, _1, verbose, logp, cat));  \
     } else ((void)0)