X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/281f8c29e177852dcb1591fc31e363e1176857c8..bdf403e0063eb38e871c0aefdc49a9bf083c316c:/src/msg/msg_vm.c diff --git a/src/msg/msg_vm.c b/src/msg/msg_vm.c index 2f773b2014..095480a18d 100644 --- a/src/msg/msg_vm.c +++ b/src/msg/msg_vm.c @@ -21,6 +21,7 @@ #include "msg_private.h" #include "xbt/sysdep.h" #include "xbt/log.h" +#include "simgrid/platf.h" XBT_LOG_NEW_DEFAULT_SUBCATEGORY(msg_vm, msg, "Cloud-oriented parts of the MSG API"); @@ -57,7 +58,7 @@ xbt_dict_t MSG_vm_get_properties(msg_vm_t vm) /** \ingroup m_host_management * \brief Change the value of a given host property * - * \param host a host + * \param vm a vm * \param name a property name * \param value what to change the property to * \param free_ctn the freeing function to use to kill the value on need @@ -798,7 +799,7 @@ const double alpha = 0.22L * 1.0E8 / (80L * 1024 * 1024); static void send_migration_data(const char *vm_name, const char *src_pm_name, const char *dst_pm_name, - double size, char *mbox, int stage, int stage2_round, double mig_speed, double xfer_cpu_overhead) + sg_size_t size, char *mbox, int stage, int stage2_round, double mig_speed, double xfer_cpu_overhead) { char *task_name = get_mig_task_name(vm_name, src_pm_name, dst_pm_name, stage); msg_task_t task = MSG_task_create(task_name, 0, size, NULL); @@ -833,9 +834,9 @@ static void send_migration_data(const char *vm_name, const char *src_pm_name, co if (stage == 2){ - XBT_DEBUG("mig-stage%d.%d: sent %f duration %f actual_speed %f (target %f) cpu %f", stage, stage2_round, size, duration, actual_speed, mig_speed, cpu_utilization);} + XBT_DEBUG("mig-stage%d.%d: sent %llu duration %f actual_speed %f (target %f) cpu %f", stage, stage2_round, size, duration, actual_speed, mig_speed, cpu_utilization);} else{ - XBT_DEBUG("mig-stage%d: sent %f duration %f actual_speed %f (target %f) cpu %f", stage, size, duration, actual_speed, mig_speed, cpu_utilization); + XBT_DEBUG("mig-stage%d: sent %llu duration %f actual_speed %f (target %f) cpu %f", stage, size, duration, actual_speed, mig_speed, cpu_utilization); } xbt_free(task_name); @@ -871,25 +872,24 @@ static double get_updated_size(double computed, double dp_rate, double dp_cap) } static double send_stage1(msg_host_t vm, const char *src_pm_name, const char *dst_pm_name, - long ramsize, double mig_speed, double xfer_cpu_overhead, double dp_rate, double dp_cap, double dpt_cpu_overhead) + sg_size_t ramsize, double mig_speed, double xfer_cpu_overhead, double dp_rate, double dp_cap, double dpt_cpu_overhead) { const char *vm_name = MSG_host_get_name(vm); char *mbox = get_mig_mbox_src_dst(vm_name, src_pm_name, dst_pm_name); - // const long chunksize = 1024 * 1024 * 100; - const long chunksize = 1024L * 1024 * 100000; - long remaining = ramsize; + // const long chunksize = (sg_size_t)1024 * 1024 * 100; + const sg_size_t chunksize = (sg_size_t)1024 * 1024 * 100000; + sg_size_t remaining = ramsize; double computed_total = 0; while (remaining > 0) { - long datasize = chunksize; + sg_size_t datasize = chunksize; if (remaining < chunksize) datasize = remaining; remaining -= datasize; send_migration_data(vm_name, src_pm_name, dst_pm_name, datasize, mbox, 1, 0, mig_speed, xfer_cpu_overhead); - double computed = lookup_computed_flop_counts(vm, 1, 0); computed_total += computed; @@ -900,7 +900,7 @@ static double send_stage1(msg_host_t vm, const char *src_pm_name, const char *ds // launch_deferred_exec_process(vm, overhead, 10000); // } } - + xbt_free(mbox); return computed_total; } @@ -928,7 +928,7 @@ static int migration_tx_fun(int argc, char *argv[]) s_ws_params_t params; simcall_host_get_params(vm, ¶ms); - const long ramsize = params.ramsize; + const sg_size_t ramsize = params.ramsize; const long devsize = params.devsize; const int skip_stage1 = params.skip_stage1; const int skip_stage2 = params.skip_stage2; @@ -974,7 +974,7 @@ static int migration_tx_fun(int argc, char *argv[]) double clock_post_send = MSG_get_clock(); double bandwidth = ramsize / (clock_post_send - clock_prev_send); threshold = get_threshold_value(bandwidth, max_downtime); - XBT_INFO("actual banwdidth %f, threshold %f", bandwidth / 1024 / 1024, threshold); + XBT_INFO("actual banwdidth %f (MB/s), threshold %f", bandwidth / 1024 / 1024, threshold); } @@ -1097,11 +1097,13 @@ static void do_migration(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm) { msg_task_t task = NULL; msg_error_t ret = MSG_task_recv(&task, mbox_ctl); + xbt_assert(ret == MSG_OK); char *expected_task_name = get_mig_task_name(sg_host_name(vm), sg_host_name(src_pm), sg_host_name(dst_pm), 4); xbt_assert(strcmp(task->name, expected_task_name) == 0); xbt_free(expected_task_name); + MSG_task_destroy(task); } xbt_free(mbox_ctl); @@ -1243,11 +1245,11 @@ msg_host_t MSG_vm_get_pm(msg_vm_t vm) * For example, * On PM0, there are Task1 and VM0. * On VM0, there is Task2. - * Now we bound 75% to Task1@PM0 and bound 25% to Task2@VM0. + * Now we bound 75% to Task1\@PM0 and bound 25% to Task2\@VM0. * Then, - * Task1@PM0 gets 50%. - * Task2@VM0 gets 25%. - * This is NOT 75% for Task1@PM0 and 25% for Task2@VM0, respectively. + * Task1\@PM0 gets 50%. + * Task2\@VM0 gets 25%. + * This is NOT 75% for Task1\@PM0 and 25% for Task2\@VM0, respectively. * * This is because a VM has the dummy CPU action in the PM layer. Putting a * task on the VM does not affect the bound of the dummy CPU action. The bound