msg_vm_t MSG_vm_get_by_name(const char *name)
{
- return MSG_get_host_by_name(name);
+ return MSG_get_host_by_name(name);
}
/** \ingroup m_vm_management
{
/* make sure the VM of the same name does not exit */
{
- void *ind_host_tmp = xbt_lib_get_elm_or_null(host_lib, name);
- if (ind_host_tmp) {
+ xbt_dictelm_t ind_host_tmp = xbt_lib_get_elm_or_null(host_lib, name);
+ if (ind_host_tmp && xbt_lib_get_level(ind_host_tmp, SIMIX_HOST_LEVEL) != NULL) {
XBT_ERROR("host %s already exits", name);
return NULL;
}
XBT_DEBUG("A new VM (%s) has been created", name);
- #ifdef HAVE_TRACING
TRACE_msg_vm_create(name, ind_pm);
- #endif
return ind_vm;
}
__MSG_host_destroy(vm);
- #ifdef HAVE_TRACING
TRACE_msg_vm_end(vm);
- #endif
}
{
simcall_vm_start(vm);
- #ifdef HAVE_TRACING
TRACE_msg_vm_start(vm);
- #endif
}
/* msg_vm_t equals to msg_host_t */
simcall_vm_shutdown(vm);
- // #ifdef HAVE_TRACING
// TRACE_msg_vm_(vm);
- // #endif
}
return bprintf("__task_mig_stage%d:%s(%s-%s)", stage, vm_name, src_pm_name, dst_pm_name);
}
-static void launch_deferred_exec_process(msg_host_t host, double computation, double prio);
-
struct migration_session {
msg_vm_t vm;
{
XBT_DEBUG("mig: rx_start");
+ // The structure has been created in the do_migration function and should only be freed in the same place ;)
struct migration_session *ms = MSG_process_get_data(MSG_process_self());
-
s_ws_params_t params;
simcall_host_get_params(ms->vm, ¶ms);
- const double xfer_cpu_overhead = params.xfer_cpu_overhead;
int need_exit = 0;
char *finalize_task_name = get_mig_task_name(ms->vm, ms->src_pm, ms->dst_pm, 3);
+ int ret = 0;
for (;;) {
msg_task_t task = NULL;
- MSG_task_recv(&task, ms->mbox);
+ ret = MSG_task_recv(&task, ms->mbox);
{
- double received = MSG_task_get_data_size(task);
- /* TODO: clean up */
- // const double alpha = 0.22L * 1.0E8 / (80L * 1024 * 1024);
- launch_deferred_exec_process(ms->vm, received * xfer_cpu_overhead, 1);
+ if (ret != MSG_OK) {
+ // An error occured, clean the code and return
+ // The owner did not change, hence the task should be only destroyed on the other side
+ xbt_free(finalize_task_name);
+ return 0;
+ }
}
if (strcmp(task->name, finalize_task_name) == 0)
break;
}
-
- /* deinstall the current affinity setting */
- simcall_vm_set_affinity(ms->vm, ms->src_pm, 0);
-
- simcall_vm_migrate(ms->vm, ms->dst_pm);
- simcall_vm_resume(ms->vm);
+ // Here Stage 1, 2 and 3 have been performed.
+ // Hence complete the migration
+
+ // Copy the reference to the vm (if SRC crashes now, do_migration will free ms)
+ // This is clearly ugly but I (Adrien) need more time to do something cleaner (actually we should copy the whole ms structure at the begining and free it at the end of each function)
+ msg_vm_t vm = ms->vm;
+ msg_host_t src_pm = ms->src_pm;
+ msg_host_t dst_pm = ms-> dst_pm;
+ msg_host_priv_t priv = msg_host_resource_priv(vm);
+
+// // TODO: we have an issue, if the DST node is turning off during the three next calls, then the VM is in an inconsistent state
+// // I should check with Takahiro in order to make this portion of code atomic
+// /* deinstall the current affinity setting for the CPU */
+// simcall_vm_set_affinity(vm, src_pm, 0);
+//
+// /* Update the vm location */
+// simcall_vm_migrate(vm, dst_pm);
+//
+// /* Resume the VM */
+// simcall_vm_resume(vm);
+//
+ simcall_vm_migratefrom_resumeto(vm, src_pm, dst_pm);
/* install the affinity setting of the VM on the destination pm */
{
- msg_host_priv_t priv = msg_host_resource_priv(ms->vm);
- unsigned long affinity_mask = (unsigned long) xbt_dict_get_or_null_ext(priv->affinity_mask_db, (char *) ms->dst_pm, sizeof(msg_host_t));
- simcall_vm_set_affinity(ms->vm, ms->dst_pm, affinity_mask);
- XBT_INFO("set affinity(0x%04lx@%s) for %s", affinity_mask, MSG_host_get_name(ms->dst_pm), MSG_host_get_name(ms->vm));
+ unsigned long affinity_mask = (unsigned long) xbt_dict_get_or_null_ext(priv->affinity_mask_db, (char *)dst_pm, sizeof(msg_host_t));
+ simcall_vm_set_affinity(vm, dst_pm, affinity_mask);
+ XBT_DEBUG("set affinity(0x%04lx@%s) for %s", affinity_mask, MSG_host_get_name(dst_pm), MSG_host_get_name(vm));
}
{
- char *task_name = get_mig_task_name(ms->vm, ms->src_pm, ms->dst_pm, 4);
+ // Now the VM is running on the new host (the migration is completed) (even if the SRC crash)
+ msg_host_priv_t priv = msg_host_resource_priv(vm);
+ priv->is_migrating = 0;
+ XBT_DEBUG("VM(%s) moved from PM(%s) to PM(%s)", ms->vm->key, ms->src_pm->key, ms->dst_pm->key);
+ TRACE_msg_vm_change_host(ms->vm, ms->src_pm, ms->dst_pm);
+ }
+ // Inform the SRC that the migration has been correctly performed
+ {
+ char *task_name = get_mig_task_name(ms->vm, ms->src_pm, ms->dst_pm, 4);
msg_task_t task = MSG_task_create(task_name, 0, 0, NULL);
msg_error_t ret = MSG_task_send(task, ms->mbox_ctl);
- xbt_assert(ret == MSG_OK);
+ // xbt_assert(ret == MSG_OK);
+ if(ret == MSG_HOST_FAILURE){
+ // The DST has crashed, this is a problem has the VM since we are not sure whether SRC is considering that the VM has been correctly migrated on the DST node
+ // TODO What does it mean ? What should we do ?
+ MSG_task_destroy(task);
+ } else if(ret == MSG_TRANSFER_FAILURE){
+ // The SRC has crashed, this is not a problem has the VM has been correctly migrated on the DST node
+ MSG_task_destroy(task);
+ }
xbt_free(task_name);
}
xbt_dict_cursor_t cursor = NULL;
dirty_page_t dp = NULL;
xbt_dict_foreach(priv->dp_objs, cursor, key, dp) {
- double remaining = MSG_task_get_remaining_computation(dp->task);
+ double remaining = MSG_task_get_flops_amount(dp->task);
dp->prev_clock = MSG_get_clock();
dp->prev_remaining = remaining;
xbt_dict_cursor_t cursor = NULL;
dirty_page_t dp = NULL;
xbt_dict_foreach(priv->dp_objs, cursor, key, dp) {
- double remaining = MSG_task_get_remaining_computation(dp->task);
+ double remaining = MSG_task_get_flops_amount(dp->task);
- double clock = MSG_get_clock();
+ double clock = MSG_get_clock();
// total += calc_updated_pages(key, vm, dp, remaining, clock);
total += get_computed(key, vm, dp, remaining, clock);
void MSG_host_add_task(msg_host_t host, msg_task_t task)
{
msg_host_priv_t priv = msg_host_resource_priv(host);
- double remaining = MSG_task_get_remaining_computation(task);
+ double remaining = MSG_task_get_flops_amount(task);
char *key = bprintf("%s-%p", task->name, task);
dirty_page_t dp = xbt_new0(s_dirty_page, 1);
* computation has been done until now, and keep the information for the
* lookup_() function that will called soon. */
if (priv->dp_enabled) {
- double remaining = MSG_task_get_remaining_computation(task);
+ double remaining = MSG_task_get_flops_amount(task);
double clock = MSG_get_clock();
// double updated = calc_updated_pages(key, host, dp, remaining, clock);
double updated = get_computed(key, host, dp, remaining, clock);
}
-static int deferred_exec_fun(int argc, char *argv[])
-{
- xbt_assert(argc == 3);
- const char *comp_str = argv[1];
- double computaion = atof(comp_str);
- const char *prio_str = argv[2];
- double prio = atof(prio_str);
-
- msg_task_t task = MSG_task_create("__task_deferred", computaion, 0, NULL);
- // XBT_INFO("exec deferred %f", computation);
-
- /* dpt is the results of the VM activity */
- MSG_task_set_priority(task, prio);
- MSG_task_execute(task);
-
-
-
- MSG_task_destroy(task);
-
- return 0;
-}
-
-static void launch_deferred_exec_process(msg_host_t host, double computation, double prio)
-{
- char *pr_name = bprintf("__pr_deferred_exec_%s", MSG_host_get_name(host));
-
- int nargvs = 4;
- char **argv = xbt_new(char *, nargvs);
- argv[0] = pr_name;
- argv[1] = bprintf("%f", computation);
- argv[2] = bprintf("%f", prio);
- argv[3] = NULL;
-
- MSG_process_create_with_arguments(pr_name, deferred_exec_fun, NULL, host, nargvs - 1, argv);
-}
-
-
-static int task_tx_overhead_fun(int argc, char *argv[])
-{
- xbt_assert(argc == 2);
- const char *mbox = argv[1];
-
- int need_exit = 0;
-
- // XBT_INFO("start %s", mbox);
-
- for (;;) {
- msg_task_t task = NULL;
- MSG_task_recv(&task, mbox);
-
- // XBT_INFO("task->name %s", task->name);
-
- if (strcmp(task->name, "finalize_making_overhead") == 0)
- need_exit = 1;
-
- // XBT_INFO("exec");
- // MSG_task_set_priority(task, 1000000);
- MSG_task_execute(task);
- MSG_task_destroy(task);
-
- if (need_exit)
- break;
- }
-
- // XBT_INFO("bye");
-
- return 0;
-}
-
-static void start_overhead_process(msg_task_t comm_task)
-{
- char *pr_name = bprintf("__pr_task_tx_overhead_%s", MSG_task_get_name(comm_task));
- char *mbox = bprintf("__mb_task_tx_overhead_%s", MSG_task_get_name(comm_task));
-
- int nargvs = 3;
- char **argv = xbt_new(char *, nargvs);
- argv[0] = pr_name;
- argv[1] = mbox;
- argv[2] = NULL;
-
- // XBT_INFO("micro start: mbox %s", mbox);
- MSG_process_create_with_arguments(pr_name, task_tx_overhead_fun, NULL, MSG_host_self(), nargvs - 1, argv);
-}
-
-static void shutdown_overhead_process(msg_task_t comm_task)
-{
- char *mbox = bprintf("__mb_task_tx_overhead_%s", MSG_task_get_name(comm_task));
-
- msg_task_t task = MSG_task_create("finalize_making_overhead", 0, 0, NULL);
-
- // XBT_INFO("micro shutdown: mbox %s", mbox);
- msg_error_t ret = MSG_task_send(task, mbox);
- xbt_assert(ret == MSG_OK);
-
- xbt_free(mbox);
- // XBT_INFO("shutdown done");
-}
-
-static void request_overhead(msg_task_t comm_task, double computation)
-{
- char *mbox = bprintf("__mb_task_tx_overhead_%s", MSG_task_get_name(comm_task));
-
- msg_task_t task = MSG_task_create("micro", computation, 0, NULL);
-
- // XBT_INFO("req overhead");
- msg_error_t ret = MSG_task_send(task, mbox);
- xbt_assert(ret == MSG_OK);
-
- xbt_free(mbox);
-}
-
-/* alpha is (floating_operations / bytes).
- *
- * When actual migration traffic was 32 mbytes/s, we observed the CPU
- * utilization of the main thread of the Qemu process was 10 %.
- * alpha = 0.1 * C / (32 * 1024 * 1024)
- * where the CPU capacity of the PM is C flops/s.
- *
- * */
-static void task_send_bounded_with_cpu_overhead(msg_task_t comm_task, char *mbox, double mig_speed, double alpha)
-{
- const double chunk_size = 1024 * 1024 * 10;
- double remaining = MSG_task_get_data_size(comm_task);
-
- start_overhead_process(comm_task);
-
-
- while (remaining > 0) {
- double data_size = chunk_size;
- if (remaining < chunk_size)
- data_size = remaining;
-
- remaining -= data_size;
-
- // XBT_INFO("remaining %f bytes", remaining);
-
-
- double clock_sta = MSG_get_clock();
-
- /* create a micro task */
- {
- char *mtask_name = bprintf("__micro_%s", MSG_task_get_name(comm_task));
- msg_task_t mtask = MSG_task_create(mtask_name, 0, data_size, NULL);
-
- request_overhead(comm_task, data_size * alpha);
-
- msg_error_t ret = MSG_task_send(mtask, mbox);
- xbt_assert(ret == MSG_OK);
-
- xbt_free(mtask_name);
- }
-
-#if 0
- {
- /* In the real world, sending data involves small CPU computation. */
- char *mtask_name = bprintf("__micro_%s", MSG_task_get_name(comm_task));
- msg_task_t mtask = MSG_task_create(mtask_name, data_size * alpha, data_size, NULL);
- MSG_task_execute(mtask);
- MSG_task_destroy(mtask);
- xbt_free(mtask_name);
- }
-#endif
-
- /* TODO */
-
- double clock_end = MSG_get_clock();
-
-
- if (mig_speed > 0) {
- /*
- * (max bandwidth) > data_size / ((elapsed time) + time_to_sleep)
- *
- * Thus, we get
- * time_to_sleep > data_size / (max bandwidth) - (elapsed time)
- *
- * If time_to_sleep is smaller than zero, the elapsed time was too big. We
- * do not need a micro sleep.
- **/
- double time_to_sleep = data_size / mig_speed - (clock_end - clock_sta);
- if (time_to_sleep > 0)
- MSG_process_sleep(time_to_sleep);
-
-
- //XBT_INFO("duration %f", clock_end - clock_sta);
- //XBT_INFO("time_to_sleep %f", time_to_sleep);
- }
- }
-
- // XBT_INFO("%s", MSG_task_get_name(comm_task));
- shutdown_overhead_process(comm_task);
-
-}
-
-
-#if 0
-static void make_cpu_overhead_of_data_transfer(msg_task_t comm_task, double init_comm_size)
-{
- double prev_remaining = init_comm_size;
-
- for (;;) {
- double remaining = MSG_task_get_remaining_communication(comm_task);
- if (remaining == 0)
- need_exit = 1;
-
- double sent = prev_remaining - remaining;
- double comp_size = sent * overhead;
- char *comp_task_name = bprintf("__sender_overhead%s", MSG_task_get_name(comm_task));
- msg_task_t comp_task = MSG_task_create(comp_task_name, comp_size, 0, NULL);
- MSG_task_execute(comp_task);
- MSG_task_destroy(comp_task);
-
- if (need_exit)
- break;
-
- prev_remaining = remaining;
-
- }
-
- xbt_free(comp_task_name);
-}
-#endif
-
-// #define USE_MICRO_TASK 1
-
-#if 0
-// const double alpha = 0.1L * 1.0E8 / (32L * 1024 * 1024);
-// const double alpha = 0.25L * 1.0E8 / (85L * 1024 * 1024);
-// const double alpha = 0.20L * 1.0E8 / (85L * 1024 * 1024);
-// const double alpha = 0.25L * 1.0E8 / (85L * 1024 * 1024);
-// const double alpha = 0.32L * 1.0E8 / (24L * 1024 * 1024); // makes super good values for 32 mbytes/s
-//const double alpha = 0.32L * 1.0E8 / (32L * 1024 * 1024);
-// const double alpha = 0.56L * 1.0E8 / (80L * 1024 * 1024);
-////const double alpha = 0.20L * 1.0E8 / (80L * 1024 * 1024);
-// const double alpha = 0.56L * 1.0E8 / (90L * 1024 * 1024);
-// const double alpha = 0.66L * 1.0E8 / (90L * 1024 * 1024);
-// const double alpha = 0.20L * 1.0E8 / (80L * 1024 * 1024);
-
-/* CPU 22% when 80Mbyte/s */
-const double alpha = 0.22L * 1.0E8 / (80L * 1024 * 1024);
-#endif
-
-
-static void send_migration_data(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm,
- sg_size_t size, char *mbox, int stage, int stage2_round, double mig_speed, double xfer_cpu_overhead)
+static sg_size_t send_migration_data(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm,
+ sg_size_t size, char *mbox, int stage, int stage2_round, double mig_speed, double timeout)
{
+ sg_size_t sent = 0;
char *task_name = get_mig_task_name(vm, src_pm, dst_pm, stage);
msg_task_t task = MSG_task_create(task_name, 0, size, NULL);
double clock_sta = MSG_get_clock();
-#ifdef USE_MICRO_TASK
-
- task_send_bounded_with_cpu_overhead(task, mbox, mig_speed, xfer_cpu_overhead);
-
-#else
msg_error_t ret;
if (mig_speed > 0)
- ret = MSG_task_send_bounded(task, mbox, mig_speed);
+ ret = MSG_task_send_with_timeout_bounded(task, mbox, timeout, mig_speed);
else
ret = MSG_task_send(task, mbox);
- xbt_assert(ret == MSG_OK);
-#endif
-
- double clock_end = MSG_get_clock();
- double duration = clock_end - clock_sta;
- double actual_speed = size / duration;
-#ifdef USE_MICRO_TASK
- double cpu_utilization = size * xfer_cpu_overhead / duration / 1.0E8;
-#else
- double cpu_utilization = 0;
-#endif
-
-
+ xbt_free(task_name);
- if (stage == 2){
- XBT_DEBUG("mig-stage%d.%d: sent %llu duration %f actual_speed %f (target %f) cpu %f", stage, stage2_round, size, duration, actual_speed, mig_speed, cpu_utilization);}
- else{
- XBT_DEBUG("mig-stage%d: sent %llu duration %f actual_speed %f (target %f) cpu %f", stage, size, duration, actual_speed, mig_speed, cpu_utilization);
+ if (ret == MSG_OK) {
+ sent = size;
+ } else if (ret == MSG_TIMEOUT) {
+ sg_size_t remaining = MSG_task_get_remaining_communication(task);
+ sent = size - remaining;
+ XBT_INFO("timeout (%lf s) in sending_migration_data, remaining %llu bytes of %llu",
+ timeout, remaining, size);
}
- xbt_free(task_name);
+ /* FIXME: why try-and-catch is used here? */
+ if(ret == MSG_HOST_FAILURE){
+ //XBT_INFO("SRC host failed during migration of %s (stage %d)", sg_host_name(vm), stage);
+ MSG_task_destroy(task);
+ THROWF(host_error, 0, "SRC host failed during migration of %s (stage %d)", sg_host_name(vm), stage);
+ }else if(ret == MSG_TRANSFER_FAILURE){
+ //XBT_INFO("DST host failed during migration of %s (stage %d)", sg_host_name(vm), stage);
+ MSG_task_destroy(task);
+ THROWF(host_error, 0, "DST host failed during migration of %s (stage %d)", sg_host_name(vm), stage);
+ }
+ double clock_end = MSG_get_clock();
+ double duration = clock_end - clock_sta;
+ double actual_speed = size / duration;
+ if (stage == 2)
+ XBT_DEBUG("mig-stage%d.%d: sent %llu duration %f actual_speed %f (target %f)", stage, stage2_round, size, duration, actual_speed, mig_speed);
+ else
+ XBT_DEBUG("mig-stage%d: sent %llu duration %f actual_speed %f (target %f)", stage, size, duration, actual_speed, mig_speed);
-#ifdef USE_MICRO_TASK
- /* The name of a micro task starts with __micro, which does not match the
- * special name that finalizes the receiver loop. Thus, we send the special task.
- **/
- {
- if (stage == 3) {
- char *task_name = get_mig_task_name(vm_name, src_pm_name, dst_pm_name, stage);
- msg_task_t task = MSG_task_create(task_name, 0, 0, NULL);
- msg_error_t ret = MSG_task_send(task, mbox);
- xbt_assert(ret == MSG_OK);
- xbt_free(task_name);
- }
- }
-#endif
+ return sent;
}
-static double get_updated_size(double computed, double dp_rate, double dp_cap)
+static sg_size_t get_updated_size(double computed, double dp_rate, double dp_cap)
{
double updated_size = computed * dp_rate;
XBT_DEBUG("updated_size %f dp_rate %f", updated_size, dp_rate);
updated_size = dp_cap;
}
- return updated_size;
+ return (sg_size_t) updated_size;
}
static double send_stage1(struct migration_session *ms,
- sg_size_t ramsize, double mig_speed, double xfer_cpu_overhead, double dp_rate, double dp_cap, double dpt_cpu_overhead)
+ sg_size_t ramsize, double mig_speed, double dp_rate, double dp_cap)
{
// const long chunksize = (sg_size_t)1024 * 1024 * 100;
datasize = remaining;
remaining -= datasize;
-
- send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, datasize, ms->mbox, 1, 0, mig_speed, xfer_cpu_overhead);
+ send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, datasize, ms->mbox, 1, 0, mig_speed, -1);
double computed = lookup_computed_flop_counts(ms->vm, 1, 0);
computed_total += computed;
-
- // {
- // double updated_size = get_updated_size(computed, dp_rate, dp_cap);
-
- // double overhead = dpt_cpu_overhead * updated_size;
- // launch_deferred_exec_process(vm, overhead, 10000);
- // }
}
return computed_total;
static double get_threshold_value(double bandwidth, double max_downtime)
{
- /* This value assumes the network link is 1Gbps. */
- // double threshold = max_downtime * 125 * 1024 * 1024;
- double threshold = max_downtime * bandwidth;
-
- return threshold;
+ return max_downtime * bandwidth;
}
static int migration_tx_fun(int argc, char *argv[])
{
XBT_DEBUG("mig: tx_start");
+ // Note that the ms structure has been allocated in do_migration and hence should be freed in the same function ;)
struct migration_session *ms = MSG_process_get_data(MSG_process_self());
s_ws_params_t params;
const sg_size_t ramsize = params.ramsize;
const sg_size_t devsize = params.devsize;
const int skip_stage1 = params.skip_stage1;
- const int skip_stage2 = params.skip_stage2;
+ int skip_stage2 = params.skip_stage2;
const double dp_rate = params.dp_rate;
const double dp_cap = params.dp_cap;
const double mig_speed = params.mig_speed;
- const double xfer_cpu_overhead = params.xfer_cpu_overhead;
- const double dpt_cpu_overhead = params.dpt_cpu_overhead;
+ double max_downtime = params.max_downtime;
+
+ /* hard code it temporally. Fix Me */
+#define MIGRATION_TIMEOUT_DO_NOT_HARDCODE_ME 10000000.0
+ double mig_timeout = MIGRATION_TIMEOUT_DO_NOT_HARDCODE_ME;
double remaining_size = ramsize + devsize;
+ double threshold = 0.0;
+
+ /* check parameters */
+ if (ramsize == 0)
+ XBT_WARN("migrate a VM, but ramsize is zero");
- double max_downtime = params.max_downtime;
if (max_downtime == 0) {
XBT_WARN("use the default max_downtime value 30ms");
max_downtime = 0.03;
}
- double threshold = 0.00001; /* TODO: cleanup */
-
- /* setting up parameters has done */
-
-
- if (ramsize == 0)
- XBT_WARN("migrate a VM, but ramsize is zero");
-
-
- XBT_INFO("mig-stage1: remaining_size %f", remaining_size);
-
/* Stage1: send all memory pages to the destination. */
+ XBT_DEBUG("mig-stage1: remaining_size %f", remaining_size);
start_dirty_page_tracking(ms->vm);
double computed_during_stage1 = 0;
if (!skip_stage1) {
- // send_migration_data(vm_name, src_pm_name, dst_pm_name, ramsize, mbox, 1, 0, mig_speed, xfer_cpu_overhead);
-
- /* send ramsize, but split it */
double clock_prev_send = MSG_get_clock();
- computed_during_stage1 = send_stage1(ms, ramsize, mig_speed, xfer_cpu_overhead, dp_rate, dp_cap, dpt_cpu_overhead);
- remaining_size -= ramsize;
+ TRY {
+ /* At stage 1, we do not need timeout. We have to send all the memory
+ * pages even though the duration of this tranfer exceeds the timeout
+ * value. */
+ XBT_INFO("Stage 1: Gonna send %llu", ramsize);
+ sg_size_t sent = send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, ramsize, ms->mbox, 1, 0, mig_speed, -1);
+ remaining_size -= sent;
+ computed_during_stage1 = lookup_computed_flop_counts(ms->vm, 1, 0);
+
+ if (sent < ramsize) {
+ XBT_INFO("mig-stage1: timeout, force moving to stage 3");
+ skip_stage2 = 1;
+ } else if (sent > ramsize)
+ XBT_CRITICAL("bug");
+
+ } CATCH_ANONYMOUS {
+ //hostfailure (if you want to know whether this is the SRC or the DST please check directly in send_migration_data code)
+ // Stop the dirty page tracking an return (there is no memory space to release)
+ stop_dirty_page_tracking(ms->vm);
+ return 0;
+ }
double clock_post_send = MSG_get_clock();
+ mig_timeout -= (clock_post_send - clock_prev_send);
+ if (mig_timeout < 0) {
+ XBT_INFO("The duration of stage 1 exceeds the timeout value (%lf > %lf), skip stage 2",
+ (clock_post_send - clock_prev_send), MIGRATION_TIMEOUT_DO_NOT_HARDCODE_ME);
+ skip_stage2 = 1;
+ }
+
+ /* estimate bandwidth */
double bandwidth = ramsize / (clock_post_send - clock_prev_send);
threshold = get_threshold_value(bandwidth, max_downtime);
- XBT_INFO("actual bandwidth %f (MB/s), threshold %f", bandwidth / 1024 / 1024, threshold);
+ XBT_DEBUG("actual bandwidth %f (MB/s), threshold %f", bandwidth / 1024 / 1024, threshold);
}
* becomes smaller than the threshold value. */
if (skip_stage2)
goto stage3;
- if (max_downtime == 0) {
- XBT_WARN("no max_downtime parameter, skip stage2");
- goto stage3;
- }
int stage2_round = 0;
for (;;) {
- double updated_size = 0;
- if (stage2_round == 0) {
- /* just after stage1, nothing has been updated. But, we have to send the data updated during stage1 */
+ sg_size_t updated_size = 0;
+ if (stage2_round == 0) {
+ /* just after stage1, nothing has been updated. But, we have to send the
+ * data updated during stage1 */
updated_size = get_updated_size(computed_during_stage1, dp_rate, dp_cap);
} else {
double computed = lookup_computed_flop_counts(ms->vm, 2, stage2_round);
updated_size = get_updated_size(computed, dp_rate, dp_cap);
}
- XBT_INFO("mig-stage 2:%d updated_size %f computed_during_stage1 %f dp_rate %f dp_cap %f",
+ XBT_DEBUG("mig-stage 2:%d updated_size %llu computed_during_stage1 %f dp_rate %f dp_cap %f",
stage2_round, updated_size, computed_during_stage1, dp_rate, dp_cap);
- // if (stage2_round != 0) {
- // /* during stage1, we have already created overhead tasks */
- // double overhead = dpt_cpu_overhead * updated_size;
- // XBT_DEBUG("updated %f overhead %f", updated_size, overhead);
- // launch_deferred_exec_process(vm, overhead, 10000);
- // }
-
-
- {
- remaining_size += updated_size;
+ /* Check whether the remaining size is below the threshold value. If so,
+ * move to stage 3. */
+ remaining_size += updated_size;
+ XBT_DEBUG("mig-stage2.%d: remaining_size %f (%s threshold %f)", stage2_round,
+ remaining_size, (remaining_size < threshold) ? "<" : ">", threshold);
+ if (remaining_size < threshold)
+ break;
- XBT_INFO("mig-stage2.%d: remaining_size %f (%s threshold %f)", stage2_round,
- remaining_size, (remaining_size < threshold) ? "<" : ">", threshold);
-
- if (remaining_size < threshold)
- break;
- }
+ sg_size_t sent = 0;
double clock_prev_send = MSG_get_clock();
-
- send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, updated_size, ms->mbox, 2, stage2_round, mig_speed, xfer_cpu_overhead);
-
+ TRY {
+ XBT_DEBUG("Stage 2, gonna send %llu", updated_size);
+ sent = send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, updated_size, ms->mbox, 2, stage2_round, mig_speed, mig_timeout);
+ } CATCH_ANONYMOUS {
+ //hostfailure (if you want to know whether this is the SRC or the DST please check directly in send_migration_data code)
+ // Stop the dirty page tracking an return (there is no memory space to release)
+ stop_dirty_page_tracking(ms->vm);
+ return 0;
+ }
double clock_post_send = MSG_get_clock();
- double bandwidth = updated_size / (clock_post_send - clock_prev_send);
- threshold = get_threshold_value(bandwidth, max_downtime);
- XBT_INFO("actual bandwidth %f, threshold %f", bandwidth / 1024 / 1024, threshold);
+ if (sent == updated_size) {
+ /* timeout did not happen */
+ double bandwidth = updated_size / (clock_post_send - clock_prev_send);
+ threshold = get_threshold_value(bandwidth, max_downtime);
+ XBT_DEBUG("actual bandwidth %f, threshold %f", bandwidth / 1024 / 1024, threshold);
+ remaining_size -= sent;
+ stage2_round += 1;
+ mig_timeout -= (clock_post_send - clock_prev_send);
+ xbt_assert(mig_timeout > 0);
+
+ } else if (sent < updated_size) {
+ /* When timeout happens, we move to stage 3. The size of memory pages
+ * updated before timeout must be added to the remaining size. */
+ XBT_INFO("mig-stage2.%d: timeout, force moving to stage 3. sent %llu / %llu, eta %lf",
+ stage2_round, sent, updated_size, (clock_post_send - clock_prev_send));
+ remaining_size -= sent;
+ double computed = lookup_computed_flop_counts(ms->vm, 2, stage2_round);
+ updated_size = get_updated_size(computed, dp_rate, dp_cap);
+ remaining_size += updated_size;
+ break;
- remaining_size -= updated_size;
- stage2_round += 1;
+ } else
+ XBT_CRITICAL("bug");
}
stage3:
/* Stage3: stop the VM and copy the rest of states. */
- XBT_INFO("mig-stage3: remaining_size %f", remaining_size);
+ XBT_DEBUG("mig-stage3: remaining_size %f", remaining_size);
simcall_vm_suspend(ms->vm);
stop_dirty_page_tracking(ms->vm);
- send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, remaining_size, ms->mbox, 3, 0, mig_speed, xfer_cpu_overhead);
-
+ TRY {
+ XBT_DEBUG("Stage 3: Gonna send %f", remaining_size);
+ send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, remaining_size, ms->mbox, 3, 0, mig_speed, -1);
+ } CATCH_ANONYMOUS {
+ //hostfailure (if you want to know whether this is the SRC or the DST please check directly in send_migration_data code)
+ // Stop the dirty page tracking an return (there is no memory space to release)
+ simcall_vm_resume(ms->vm);
+ return 0;
+ }
+ // At that point the Migration is considered valid for the SRC node but remind that the DST side should relocate effectively the VM on the DST node.
XBT_DEBUG("mig: tx_done");
return 0;
-static void do_migration(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
+static int do_migration(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
{
struct migration_session *ms = xbt_new(struct migration_session, 1);
ms->vm = vm;
ms->dst_pm = dst_pm;
ms->mbox_ctl = get_mig_mbox_ctl(vm, src_pm, dst_pm);
ms->mbox = get_mig_mbox_src_dst(vm, src_pm, dst_pm);
+
char *pr_rx_name = get_mig_process_rx_name(vm, src_pm, dst_pm);
char *pr_tx_name = get_mig_process_tx_name(vm, src_pm, dst_pm);
- MSG_process_create(pr_rx_name, migration_rx_fun, ms, dst_pm);
- MSG_process_create(pr_tx_name, migration_tx_fun, ms, src_pm);
+// msg_process_t tx_process, rx_process;
+// MSG_process_create(pr_rx_name, migration_rx_fun, ms, dst_pm);
+// MSG_process_create(pr_tx_name, migration_tx_fun, ms, src_pm);
+#if 1
+ {
+ char **argv = xbt_new(char *, 2);
+ argv[0] = pr_rx_name;
+ argv[1] = NULL;
+/*rx_process = */ MSG_process_create_with_arguments(pr_rx_name, migration_rx_fun, ms, dst_pm, 1, argv);
+ }
+ {
+ char **argv = xbt_new(char *, 2);
+ argv[0] = pr_tx_name;
+ argv[1] = NULL;
+/* tx_process = */MSG_process_create_with_arguments(pr_tx_name, migration_tx_fun, ms, src_pm, 1, argv);
+ }
+#endif
- /* wait until the migration have finished */
+ /* wait until the migration have finished or on error has occured */
{
+ XBT_DEBUG("wait for reception of the final ACK (i.e. migration has been correctly performed");
msg_task_t task = NULL;
- msg_error_t ret = MSG_task_recv(&task, ms->mbox_ctl);
+ msg_error_t ret = MSG_TIMEOUT;
+ while (ret == MSG_TIMEOUT && MSG_host_is_on(dst_pm)) //Wait while you receive the message o
+ ret = MSG_task_receive_with_timeout(&task, ms->mbox_ctl, 4);
- xbt_assert(ret == MSG_OK);
+ xbt_free(ms->mbox_ctl);
+ xbt_free(ms->mbox);
+ xbt_free(ms);
+
+ //xbt_assert(ret == MSG_OK);
+ if(ret == MSG_HOST_FAILURE){
+ // Note that since the communication failed, the owner did not change and the task should be destroyed on the other side.
+ // Hence, just throw the execption
+ XBT_INFO("SRC crashes, throw an exception (m-control)");
+ //MSG_process_kill(tx_process); // Adrien, I made a merge on Nov 28th 2014, I'm not sure whether this line is required or not
+ return -1;
+ }
+ else if((ret == MSG_TRANSFER_FAILURE) || (ret == MSG_TIMEOUT)){ // MSG_TIMEOUT here means that MSG_host_is_avail() returned false.
+ XBT_INFO("DST crashes, throw an exception (m-control)");
+ return -2;
+ }
+
char *expected_task_name = get_mig_task_name(vm, src_pm, dst_pm, 4);
xbt_assert(strcmp(task->name, expected_task_name) == 0);
xbt_free(expected_task_name);
MSG_task_destroy(task);
+ return 0;
}
-
- xbt_free(ms->mbox_ctl);
- xbt_free(ms->mbox);
- xbt_free(ms);
}
+
+
/** @brief Migrate the VM to the given host.
* @ingroup msg_VMs
*
* physical host.
*
* The second one would be easier.
- *
+ *
*/
msg_host_t old_pm = simcall_vm_get_pm(vm);
+ if(MSG_host_is_off(old_pm))
+ THROWF(vm_error, 0, "SRC host(%s) seems off, cannot start a migration", sg_host_name(old_pm));
+
+ if(MSG_host_is_off(new_pm))
+ THROWF(vm_error, 0, "DST host(%s) seems off, cannot start a migration", sg_host_name(new_pm));
+
if (!MSG_vm_is_running(vm))
THROWF(vm_error, 0, "VM(%s) is not running", sg_host_name(vm));
msg_host_priv_t priv = msg_host_resource_priv(vm);
priv->is_migrating = 1;
- do_migration(vm, old_pm, new_pm);
-
- priv->is_migrating = 0;
-
- XBT_DEBUG("VM(%s) moved from PM(%s) to PM(%s)", vm->key, old_pm->key, new_pm->key);
+ {
+
+ int ret = do_migration(vm, old_pm, new_pm);
+ if (ret == -1){
+ priv->is_migrating = 0;
+ THROWF(host_error, 0, "SRC host failed during migration");
+ }
+ else if(ret == -2){
+ priv->is_migrating = 0;
+ THROWF(host_error, 0, "DST host failed during migration");
+ }
+ }
- #ifdef HAVE_TRACING
- TRACE_msg_vm_change_host(vm, old_pm, new_pm);
- #endif
+ // This part is done in the RX code, to handle the corner case where SRC can crash just at the end of the migration process
+ // In that case, the VM has been already assigned to the DST node.
+ //XBT_DEBUG("VM(%s) moved from PM(%s) to PM(%s)", vm->key, old_pm->key, new_pm->key);
+ //TRACE_msg_vm_change_host(vm, old_pm, new_pm);
}
XBT_DEBUG("vm_suspend done");
- #ifdef HAVE_TRACING
TRACE_msg_vm_suspend(vm);
- #endif
}
{
simcall_vm_resume(vm);
- #ifdef HAVE_TRACING
TRACE_msg_vm_resume(vm);
- #endif
}
THROWF(vm_error, 0, "VM(%s) is migrating", sg_host_name(vm));
simcall_vm_save(vm);
- #ifdef HAVE_TRACING
TRACE_msg_vm_save(vm);
- #endif
}
/** @brief Restore the execution of the VM. All processes on the VM run again.
{
simcall_vm_restore(vm);
- #ifdef HAVE_TRACING
TRACE_msg_vm_restore(vm);
- #endif
}
* On PM0, there are Task1 and VM0.
* On VM0, there is Task2.
* Now we bound 75% to Task1\@PM0 and bound 25% to Task2\@VM0.
- * Then,
+ * Then,
* Task1\@PM0 gets 50%.
* Task2\@VM0 gets 25%.
* This is NOT 75% for Task1\@PM0 and 25% for Task2\@VM0, respectively.
*/
void MSG_vm_set_bound(msg_vm_t vm, double bound)
{
- return simcall_vm_set_bound(vm, bound);
+ return simcall_vm_set_bound(vm, bound);
}
msg_host_t pm_now = MSG_vm_get_pm(vm);
if (pm_now == pm) {
- XBT_INFO("set affinity(0x%04lx@%s) for %s", mask, MSG_host_get_name(pm), MSG_host_get_name(vm));
+ XBT_DEBUG("set affinity(0x%04lx@%s) for %s", mask, MSG_host_get_name(pm), MSG_host_get_name(vm));
simcall_vm_set_affinity(vm, pm, mask);
} else
- XBT_INFO("set affinity(0x%04lx@%s) for %s (not active now)", mask, MSG_host_get_name(pm), MSG_host_get_name(vm));
+ XBT_DEBUG("set affinity(0x%04lx@%s) for %s (not active now)", mask, MSG_host_get_name(pm), MSG_host_get_name(vm));
}