]> AND Private Git Repository - loba.git/blobdiff - process.h
Logo AND Algorithmique Numérique Distribuée

Private GIT Repository
TESTS are ready to run!
[loba.git] / process.h
index aa1e032122e18545ce45d3d20ef4735ada3583a0..8a9e74c3d63da11bd185dfc525b567cf694e6b61 100644 (file)
--- a/process.h
+++ b/process.h
 #ifndef PROCESS_H
 #define PROCESS_H
 
+#define USE_UNORDERED_MAP 1
+//#undef USE_UNORDERED_MAP
+
+#include <algorithm>
+#include <functional>
+#ifdef USE_UNORDERED_MAP
+#  include <unordered_map>
+#  define MAP_TEMPLATE std::unordered_map
+#else
+#  include <map>
+#  define MAP_TEMPLATE std::map
+#endif
 #include <vector>
+#include <msg/msg.h>
 #include <xbt/log.h>
 #include "communicator.h"
+#include "misc.h"
+#include "msg_thread.h"
 #include "neighbor.h"
+#include "options.h"
+#include "synchro.h"
 
 class process {
 public:
-    process(int argc, char *argv[]);
-    ~process() { };
+    static double get_total_load_init()    { return total_load_init;    }
+    static double get_total_load_running() { return total_load_running; }
+    static double get_total_load_exit()    { return total_load_exit;    }
+
+    process(int argc, char* argv[]);
+    virtual ~process();
+
+    double get_real_load() const           { return real_load;            }
+    double get_comp_amount() const         { return acc.comp_amount;      }
+    double get_data_send_amount() const    { return acc.data_send.amount; }
+    double get_data_recv_amount() const    { return acc.data_recv.amount; }
+    unsigned get_data_send_count() const   { return acc.data_send.count;  }
+    unsigned get_data_recv_count() const   { return acc.data_recv.count;  }
+    double get_ctrl_send_amount() const    { return acc.ctrl_send.amount; }
+    double get_ctrl_recv_amount() const    { return acc.ctrl_recv.amount; }
+    unsigned get_ctrl_send_count() const   { return acc.ctrl_send.count;  }
+    unsigned get_ctrl_recv_count() const   { return acc.ctrl_recv.count;  }
+
     int run();
 
+protected:
+    typedef std::vector<neighbor> neigh_type;
+    typedef std::vector<neighbor*> pneigh_type;
+
+    pneigh_type pneigh;         // list of pointers to neighbors that
+                                // we are free to reorder
+
+    // Get and set current load, which may be real load, or expected
+    // load if opt::bookkeeping is true.
+    double get_load() const                { return expected_load; }
+
+    // The load balancing algorithm comes here...
+    virtual void load_balance();
+
+    // Register some amount of load to send to given neighbor.
+    void send(neighbor& nb, double amount);
+    void send(neighbor* nb, double amount) { send(*nb, amount); }
+
+    // Sort pneigh by applying comp to their loads
+    template <typename Compare>
+    void pneigh_sort_by_load(const Compare& comp);
+
+    // Calls neighbor::print(verbose, logp, cat) for each member of neigh.
+    void print_loads(bool verbose = false,
+                     e_xbt_log_priority_t logp = xbt_log_priority_info,
+                     xbt_log_category_t cat = _XBT_LOGV(default)) const;
+
+    // Calls neighbor::print(verbose, logp, cat) for each member of pneigh.
+    void print_loads_p(bool verbose = false,
+                       e_xbt_log_priority_t logp = xbt_log_priority_info,
+                       xbt_log_category_t cat = _XBT_LOGV(default)) const;
+
 private:
-    communicator comm;
-    std::vector<neighbor> neigh;
-    double load;
+    static double total_load_init; // sum of process loads at init
+    static double total_load_running; // sum of loads while running
+    static double total_load_exit; // sum of process loads at exit
+
+    typedef MAP_TEMPLATE<m_host_t, neighbor*> rev_neigh_type;
+    neigh_type neigh;           // list of neighbors (do not alter
+                                // after construction!)
+    rev_neigh_type rev_neigh;   // map m_host_t -> neighbor
 
-    void receive();
-    void compute();
-    void print_loads(e_xbt_log_priority_t logp = xbt_log_priority_info);
+    communicator comm;          // communicator for this process
+    int ctrl_close_pending;     // number of "close" messages to wait
+                                // on ctrl channel
+    int data_close_pending;     // number of "close" messages to wait
+                                // on data channel
+    bool close_received;        // true if we received a "close" message
+    bool finalizing;            // true when finalize() is running
+
+    unsigned lb_iter;           // counter of load-balancing iterations
+    unsigned comp_iter;         // counter of computation iterations
+
+    double prev_load_broadcast; // used to ensure that we do not send
+                                // a same information messages
+    double real_load;           // current load
+    double expected_load;       // expected load in bookkeeping mode
+    double received_load;       // load received from neighbors
+
+    mutex_t mutex;              // synchronization between threads
+    condition_t cond;
+
+    struct mesg_accounting {
+        double amount;          // sum of message size
+        unsigned count;         // number of messages
+        mesg_accounting(): amount(0.0), count(0) { }
+    };
+    struct accounting {
+        double comp_amount;        // total computing done so far (flops)
+        mesg_accounting data_send; // data messages sent
+        mesg_accounting data_recv; // data messages received
+        mesg_accounting ctrl_send; // ctrl message sent
+        mesg_accounting ctrl_recv; // ctrl message received
+        accounting(): comp_amount(0.0) { }
+    };
+    accounting acc;
+
+    void add_comp_amount(double amount) { acc.comp_amount += amount; }
+    void add_data_send_mesg(double amount) {
+        ++acc.data_send.count;
+        acc.data_send.amount += amount;
+    }
+    void add_data_recv_mesg(double amount) {
+        ++acc.data_recv.count;
+        acc.data_recv.amount += amount;
+    }
+    void add_ctrl_send_mesg(double amount) {
+        ++acc.ctrl_send.count;
+        acc.ctrl_send.amount += amount;
+    }
+    void add_ctrl_recv_mesg(double amount) {
+        ++acc.ctrl_recv.count;
+        acc.ctrl_recv.amount += amount;
+    }
+
+    // Load-balancing loop
+    msg_thread* lb_thread;
+    void load_balance_loop();
+
+    // Simulate computation loop
+    void compute_loop();
+
+    // Check if we need to stop
+    bool still_running();
+
+    // Returns the sum of "to_send" for all neighbors.
+    double get_sum_of_to_send() const;
+
+    // Compute load_to_send (for data_send), subject to the execution parameters
+    static double compute_load_to_send(double desired);
+
+    // Send procedures
+    void ctrl_send(neighbor& nb);
+    void data_send(neighbor& nb);
+    void ctrl_close(neighbor& nb);
+    void data_close(neighbor& nb);
+
+    // Receive procedure
+    // Parameter "timeout" may be 0 for non-blocking operation, -1 for
+    // infinite waiting, or any positive timeout.
+    void ctrl_receive(double timeout);
+    void data_receive(double timeout);
+    void handle_message(message* msg, m_host_t from);
 };
 
+template <typename Compare>
+void process::pneigh_sort_by_load(const Compare& comp)
+{
+    using std::placeholders::_1;
+    using std::placeholders::_2;
+    std::sort(pneigh.begin(), pneigh.end(),
+              std::bind(comp,
+                        std::bind(&neighbor::get_load, _1),
+                        std::bind(&neighbor::get_load, _2)));
+}
+
 #endif // !PROCESS_H
 
 // Local variables: