1 /* Copyright (c) 2012-2014. The SimGrid Team.
2 * All rights reserved. */
4 /* This program is free software; you can redistribute it and/or modify it
5 * under the terms of the license (GNU LGPL) which comes with this package. */
8 * 1. add the support of trace
9 * 2. use parallel tasks to simulate CPU overhead and remove the very
10 * experimental code generating micro computation tasks
15 #include "msg_private.h"
16 #include "xbt/sysdep.h"
18 #include "simgrid/platf.h"
20 XBT_LOG_NEW_DEFAULT_SUBCATEGORY(msg_vm, msg,
21 "Cloud-oriented parts of the MSG API");
24 /* **** ******** GENERAL ********* **** */
26 /** \ingroup m_vm_management
27 * \brief Returns the value of a given vm property
30 * \param name a property name
31 * \return value of a property (or NULL if property not set)
34 const char *MSG_vm_get_property_value(msg_vm_t vm, const char *name)
36 return MSG_host_get_property_value(vm, name);
39 /** \ingroup m_vm_management
40 * \brief Returns a xbt_dict_t consisting of the list of properties assigned to this host
43 * \return a dict containing the properties
45 xbt_dict_t MSG_vm_get_properties(msg_vm_t vm)
47 xbt_assert((vm != NULL), "Invalid parameters (vm is NULL)");
49 return (simcall_host_get_properties(vm));
52 /** \ingroup m_host_management
53 * \brief Change the value of a given host property
56 * \param name a property name
57 * \param value what to change the property to
58 * \param free_ctn the freeing function to use to kill the value on need
60 void MSG_vm_set_property_value(msg_vm_t vm, const char *name, void *value, void_f_pvoid_t free_ctn)
62 xbt_dict_set(MSG_host_get_properties(vm), name, value, free_ctn);
65 /** \ingroup msg_vm_management
66 * \brief Finds a msg_vm_t using its name.
68 * This is a name directory service
69 * \param name the name of a vm.
70 * \return the corresponding vm
72 * Please note that a VM is a specific host. Hence, you should give a different name
76 msg_vm_t MSG_vm_get_by_name(const char *name)
78 return MSG_get_host_by_name(name);
81 /** \ingroup m_vm_management
83 * \brief Return the name of the #msg_host_t.
85 * This functions checks whether \a host is a valid pointer or not and return
88 const char *MSG_vm_get_name(msg_vm_t vm)
90 return MSG_host_get_name(vm);
94 /* **** Check state of a VM **** */
95 static inline int __MSG_vm_is_state(msg_vm_t vm, e_surf_vm_state_t state)
97 return simcall_vm_get_state(vm) == state;
100 /** @brief Returns whether the given VM has just created, not running.
103 int MSG_vm_is_created(msg_vm_t vm)
105 return __MSG_vm_is_state(vm, SURF_VM_STATE_CREATED);
108 /** @brief Returns whether the given VM is currently running
111 int MSG_vm_is_running(msg_vm_t vm)
113 return __MSG_vm_is_state(vm, SURF_VM_STATE_RUNNING);
116 /** @brief Returns whether the given VM is currently migrating
119 int MSG_vm_is_migrating(msg_vm_t vm)
121 msg_host_priv_t priv = msg_host_resource_priv(vm);
122 return priv->is_migrating;
125 /** @brief Returns whether the given VM is currently suspended, not running.
128 int MSG_vm_is_suspended(msg_vm_t vm)
130 return __MSG_vm_is_state(vm, SURF_VM_STATE_SUSPENDED);
133 /** @brief Returns whether the given VM is being saved (FIXME: live saving or not?).
136 int MSG_vm_is_saving(msg_vm_t vm)
138 return __MSG_vm_is_state(vm, SURF_VM_STATE_SAVING);
141 /** @brief Returns whether the given VM has been saved, not running.
144 int MSG_vm_is_saved(msg_vm_t vm)
146 return __MSG_vm_is_state(vm, SURF_VM_STATE_SAVED);
149 /** @brief Returns whether the given VM is being restored, not running.
152 int MSG_vm_is_restoring(msg_vm_t vm)
154 return __MSG_vm_is_state(vm, SURF_VM_STATE_RESTORING);
159 /* ------------------------------------------------------------------------- */
160 /* ------------------------------------------------------------------------- */
162 /* **** ******** MSG vm actions ********* **** */
164 /** @brief Create a new VM with specified parameters.
166 * All parameters are in MBytes
169 msg_vm_t MSG_vm_create(msg_host_t ind_pm, const char *name,
170 int ncpus, int ramsize,
171 int net_cap, char *disk_path, int disksize,
172 int mig_netspeed, int dp_intensity)
174 /* For the moment, intensity_rate is the percentage against the migration
176 double host_speed = MSG_get_host_speed(ind_pm);
177 double update_speed = ((double)dp_intensity/100) * mig_netspeed;
179 msg_vm_t vm = MSG_vm_create_core(ind_pm, name);
180 s_ws_params_t params;
181 memset(¶ms, 0, sizeof(params));
182 params.ramsize = (sg_size_t)ramsize * 1024 * 1024;
183 //params.overcommit = 0;
185 params.skip_stage2 = 0;
186 params.max_downtime = 0.03;
187 params.dp_rate = (update_speed * 1024 * 1024) / host_speed;
188 params.dp_cap = params.ramsize * 0.9; // assume working set memory is 90% of ramsize
189 params.mig_speed = (double)mig_netspeed * 1024 * 1024; // mig_speed
191 //XBT_INFO("dp rate %f migspeed : %f intensity mem : %d, updatespeed %f, hostspeed %f",params.dp_rate, params.mig_speed, dp_intensity, update_speed, host_speed);
192 simcall_host_set_params(vm, ¶ms);
198 /** @brief Create a new VM object. The VM is not yet started. The resource of the VM is allocated upon MSG_vm_start().
201 * A VM is treated as a host. The name of the VM must be unique among all hosts.
203 msg_vm_t MSG_vm_create_core(msg_host_t ind_pm, const char *name)
205 /* make sure the VM of the same name does not exit */
207 void *ind_host_tmp = xbt_lib_get_elm_or_null(host_lib, name);
209 XBT_ERROR("host %s already exits", name);
214 /* Note: ind_vm and vm_workstation point to the same elm object. */
215 msg_vm_t ind_vm = NULL;
216 void *ind_vm_workstation = NULL;
218 /* Ask the SIMIX layer to create the surf vm resource */
219 ind_vm_workstation = simcall_vm_create(name, ind_pm);
220 ind_vm = (msg_vm_t) __MSG_host_create(ind_vm_workstation);
222 XBT_DEBUG("A new VM (%s) has been created", name);
225 TRACE_msg_vm_create(name, ind_pm);
231 /** @brief Destroy a VM. Destroy the VM object from the simulation.
234 void MSG_vm_destroy(msg_vm_t vm)
236 if (MSG_vm_is_migrating(vm))
237 THROWF(vm_error, 0, "VM(%s) is migrating", sg_host_name(vm));
239 /* First, terminate all processes on the VM if necessary */
240 if (MSG_vm_is_running(vm))
241 simcall_vm_shutdown(vm);
243 if (!MSG_vm_is_created(vm)) {
244 XBT_CRITICAL("shutdown the given VM before destroying it");
248 /* Then, destroy the VM object */
249 simcall_vm_destroy(vm);
251 __MSG_host_destroy(vm);
254 TRACE_msg_vm_end(vm);
259 /** @brief Start a vm (i.e., boot the guest operating system)
262 * If the VM cannot be started, an exception is generated.
265 void MSG_vm_start(msg_vm_t vm)
267 simcall_vm_start(vm);
270 TRACE_msg_vm_start(vm);
276 /** @brief Immediately kills all processes within the given VM. Any memory that they allocated will be leaked.
279 * FIXME: No extra delay occurs. If you want to simulate this too, you want to
280 * use a #MSG_process_sleep() or something. I'm not quite sure.
282 void MSG_vm_shutdown(msg_vm_t vm)
284 /* msg_vm_t equals to msg_host_t */
285 simcall_vm_shutdown(vm);
287 // #ifdef HAVE_TRACING
288 // TRACE_msg_vm_(vm);
294 /* We have two mailboxes. mbox is used to transfer migration data between
295 * source and destination PMs. mbox_ctl is used to detect the completion of a
296 * migration. The names of these mailboxes must not conflict with others. */
297 static inline char *get_mig_mbox_src_dst(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
299 char *vm_name = sg_host_name(vm);
300 char *src_pm_name = sg_host_name(src_pm);
301 char *dst_pm_name = sg_host_name(dst_pm);
303 return bprintf("__mbox_mig_src_dst:%s(%s-%s)", vm_name, src_pm_name, dst_pm_name);
306 static inline char *get_mig_mbox_ctl(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
308 char *vm_name = sg_host_name(vm);
309 char *src_pm_name = sg_host_name(src_pm);
310 char *dst_pm_name = sg_host_name(dst_pm);
312 return bprintf("__mbox_mig_ctl:%s(%s-%s)", vm_name, src_pm_name, dst_pm_name);
315 static inline char *get_mig_process_tx_name(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
317 char *vm_name = sg_host_name(vm);
318 char *src_pm_name = sg_host_name(src_pm);
319 char *dst_pm_name = sg_host_name(dst_pm);
321 return bprintf("__pr_mig_tx:%s(%s-%s)", vm_name, src_pm_name, dst_pm_name);
324 static inline char *get_mig_process_rx_name(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
326 char *vm_name = sg_host_name(vm);
327 char *src_pm_name = sg_host_name(src_pm);
328 char *dst_pm_name = sg_host_name(dst_pm);
330 return bprintf("__pr_mig_rx:%s(%s-%s)", vm_name, src_pm_name, dst_pm_name);
333 static inline char *get_mig_task_name(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm, int stage)
335 char *vm_name = sg_host_name(vm);
336 char *src_pm_name = sg_host_name(src_pm);
337 char *dst_pm_name = sg_host_name(dst_pm);
339 return bprintf("__task_mig_stage%d:%s(%s-%s)", stage, vm_name, src_pm_name, dst_pm_name);
342 static void launch_deferred_exec_process(msg_host_t host, double computation, double prio);
345 struct migration_session {
350 /* The miration_rx process uses mbox_ctl to let the caller of do_migration()
351 * know the completion of the migration. */
353 /* The migration_rx and migration_tx processes use mbox to transfer migration
359 static int migration_rx_fun(int argc, char *argv[])
361 XBT_DEBUG("mig: rx_start");
363 struct migration_session *ms = MSG_process_get_data(MSG_process_self());
366 s_ws_params_t params;
367 simcall_host_get_params(ms->vm, ¶ms);
368 const double xfer_cpu_overhead = params.xfer_cpu_overhead;
372 char *finalize_task_name = get_mig_task_name(ms->vm, ms->src_pm, ms->dst_pm, 3);
375 msg_task_t task = NULL;
376 MSG_task_recv(&task, ms->mbox);
378 double received = MSG_task_get_data_size(task);
380 // const double alpha = 0.22L * 1.0E8 / (80L * 1024 * 1024);
381 launch_deferred_exec_process(ms->vm, received * xfer_cpu_overhead, 1);
384 if (strcmp(task->name, finalize_task_name) == 0)
387 MSG_task_destroy(task);
394 /* deinstall the current affinity setting for the CPU */
395 simcall_vm_set_affinity(ms->vm, ms->src_pm, 0);
397 /* Update the vm location */
398 simcall_vm_migrate(ms->vm, ms->dst_pm);
401 simcall_vm_resume(ms->vm);
403 /* install the affinity setting of the VM on the destination pm */
405 msg_host_priv_t priv = msg_host_resource_priv(ms->vm);
407 unsigned long affinity_mask = (unsigned long) xbt_dict_get_or_null_ext(priv->affinity_mask_db, (char *) ms->dst_pm, sizeof(msg_host_t));
408 simcall_vm_set_affinity(ms->vm, ms->dst_pm, affinity_mask);
409 XBT_INFO("set affinity(0x%04lx@%s) for %s", affinity_mask, MSG_host_get_name(ms->dst_pm), MSG_host_get_name(ms->vm));
413 char *task_name = get_mig_task_name(ms->vm, ms->src_pm, ms->dst_pm, 4);
415 msg_task_t task = MSG_task_create(task_name, 0, 0, NULL);
416 msg_error_t ret = MSG_task_send(task, ms->mbox_ctl);
417 xbt_assert(ret == MSG_OK);
423 xbt_free(finalize_task_name);
425 XBT_DEBUG("mig: rx_done");
430 static void reset_dirty_pages(msg_vm_t vm)
432 msg_host_priv_t priv = msg_host_resource_priv(vm);
435 xbt_dict_cursor_t cursor = NULL;
436 dirty_page_t dp = NULL;
437 xbt_dict_foreach(priv->dp_objs, cursor, key, dp) {
438 double remaining = MSG_task_get_remaining_computation(dp->task);
439 dp->prev_clock = MSG_get_clock();
440 dp->prev_remaining = remaining;
442 // XBT_INFO("%s@%s remaining %f", key, sg_host_name(vm), remaining);
446 static void start_dirty_page_tracking(msg_vm_t vm)
448 msg_host_priv_t priv = msg_host_resource_priv(vm);
449 priv->dp_enabled = 1;
451 reset_dirty_pages(vm);
454 static void stop_dirty_page_tracking(msg_vm_t vm)
456 msg_host_priv_t priv = msg_host_resource_priv(vm);
457 priv->dp_enabled = 0;
461 /* It might be natural that we define dp_rate for each task. But, we will also
462 * have to care about how each task behavior affects the memory update behavior
463 * at the operating system level. It may not be easy to model it with a simple algorithm. */
464 double calc_updated_pages(char *key, msg_vm_t vm, dirty_page_t dp, double remaining, double clock)
466 double computed = dp->prev_remaining - remaining;
467 double duration = clock - dp->prev_clock;
468 double updated = dp->task->dp_rate * computed;
470 XBT_INFO("%s@%s: computated %f ops (remaining %f -> %f) in %f secs (%f -> %f)",
471 key, sg_host_name(vm), computed, dp->prev_remaining, remaining, duration, dp->prev_clock, clock);
472 XBT_INFO("%s@%s: updated %f bytes, %f Mbytes/s",
473 key, sg_host_name(vm), updated, updated / duration / 1000 / 1000);
479 static double get_computed(char *key, msg_vm_t vm, dirty_page_t dp, double remaining, double clock)
481 double computed = dp->prev_remaining - remaining;
482 double duration = clock - dp->prev_clock;
484 XBT_DEBUG("%s@%s: computed %f ops (remaining %f -> %f) in %f secs (%f -> %f)",
485 key, sg_host_name(vm), computed, dp->prev_remaining, remaining, duration, dp->prev_clock, clock);
490 static double lookup_computed_flop_counts(msg_vm_t vm, int stage_for_fancy_debug, int stage2_round_for_fancy_debug)
492 msg_host_priv_t priv = msg_host_resource_priv(vm);
496 xbt_dict_cursor_t cursor = NULL;
497 dirty_page_t dp = NULL;
498 xbt_dict_foreach(priv->dp_objs, cursor, key, dp) {
499 double remaining = MSG_task_get_remaining_computation(dp->task);
501 double clock = MSG_get_clock();
503 // total += calc_updated_pages(key, vm, dp, remaining, clock);
504 total += get_computed(key, vm, dp, remaining, clock);
506 dp->prev_remaining = remaining;
507 dp->prev_clock = clock;
510 total += priv->dp_updated_by_deleted_tasks;
512 XBT_DEBUG("mig-stage%d.%d: computed %f flop_counts (including %f by deleted tasks)",
513 stage_for_fancy_debug,
514 stage2_round_for_fancy_debug,
515 total, priv->dp_updated_by_deleted_tasks);
519 priv->dp_updated_by_deleted_tasks = 0;
525 // TODO Is this code redundant with the information provided by
526 // msg_process_t MSG_process_create(const char *name, xbt_main_func_t code, void *data, msg_host_t host)
527 void MSG_host_add_task(msg_host_t host, msg_task_t task)
529 msg_host_priv_t priv = msg_host_resource_priv(host);
530 double remaining = MSG_task_get_remaining_computation(task);
531 char *key = bprintf("%s-%p", task->name, task);
533 dirty_page_t dp = xbt_new0(s_dirty_page, 1);
536 /* It should be okay that we add a task onto a migrating VM. */
537 if (priv->dp_enabled) {
538 dp->prev_clock = MSG_get_clock();
539 dp->prev_remaining = remaining;
542 xbt_assert(xbt_dict_get_or_null(priv->dp_objs, key) == NULL);
543 xbt_dict_set(priv->dp_objs, key, dp, NULL);
544 XBT_DEBUG("add %s on %s (remaining %f, dp_enabled %d)", key, sg_host_name(host), remaining, priv->dp_enabled);
549 void MSG_host_del_task(msg_host_t host, msg_task_t task)
551 msg_host_priv_t priv = msg_host_resource_priv(host);
553 char *key = bprintf("%s-%p", task->name, task);
555 dirty_page_t dp = xbt_dict_get_or_null(priv->dp_objs, key);
556 xbt_assert(dp->task == task);
558 /* If we are in the middle of dirty page tracking, we record how much
559 * computation has been done until now, and keep the information for the
560 * lookup_() function that will called soon. */
561 if (priv->dp_enabled) {
562 double remaining = MSG_task_get_remaining_computation(task);
563 double clock = MSG_get_clock();
564 // double updated = calc_updated_pages(key, host, dp, remaining, clock);
565 double updated = get_computed(key, host, dp, remaining, clock);
567 priv->dp_updated_by_deleted_tasks += updated;
570 xbt_dict_remove(priv->dp_objs, key);
573 XBT_DEBUG("del %s on %s", key, sg_host_name(host));
579 static int deferred_exec_fun(int argc, char *argv[])
581 xbt_assert(argc == 3);
582 const char *comp_str = argv[1];
583 double computaion = atof(comp_str);
584 const char *prio_str = argv[2];
585 double prio = atof(prio_str);
587 msg_task_t task = MSG_task_create("__task_deferred", computaion, 0, NULL);
588 // XBT_INFO("exec deferred %f", computation);
590 /* dpt is the results of the VM activity */
591 MSG_task_set_priority(task, prio);
592 MSG_task_execute(task);
596 MSG_task_destroy(task);
601 static void launch_deferred_exec_process(msg_host_t host, double computation, double prio)
603 char *pr_name = bprintf("__pr_deferred_exec_%s", MSG_host_get_name(host));
606 char **argv = xbt_new(char *, nargvs);
608 argv[1] = bprintf("%f", computation);
609 argv[2] = bprintf("%f", prio);
612 MSG_process_create_with_arguments(pr_name, deferred_exec_fun, NULL, host, nargvs - 1, argv);
616 static int task_tx_overhead_fun(int argc, char *argv[])
618 xbt_assert(argc == 2);
619 const char *mbox = argv[1];
623 // XBT_INFO("start %s", mbox);
626 msg_task_t task = NULL;
627 MSG_task_recv(&task, mbox);
629 // XBT_INFO("task->name %s", task->name);
631 if (strcmp(task->name, "finalize_making_overhead") == 0)
635 // MSG_task_set_priority(task, 1000000);
636 MSG_task_execute(task);
637 MSG_task_destroy(task);
648 static void start_overhead_process(msg_task_t comm_task)
650 char *pr_name = bprintf("__pr_task_tx_overhead_%s", MSG_task_get_name(comm_task));
651 char *mbox = bprintf("__mb_task_tx_overhead_%s", MSG_task_get_name(comm_task));
654 char **argv = xbt_new(char *, nargvs);
659 // XBT_INFO("micro start: mbox %s", mbox);
660 MSG_process_create_with_arguments(pr_name, task_tx_overhead_fun, NULL, MSG_host_self(), nargvs - 1, argv);
663 static void shutdown_overhead_process(msg_task_t comm_task)
665 char *mbox = bprintf("__mb_task_tx_overhead_%s", MSG_task_get_name(comm_task));
667 msg_task_t task = MSG_task_create("finalize_making_overhead", 0, 0, NULL);
669 // XBT_INFO("micro shutdown: mbox %s", mbox);
670 msg_error_t ret = MSG_task_send(task, mbox);
671 xbt_assert(ret == MSG_OK);
674 // XBT_INFO("shutdown done");
677 static void request_overhead(msg_task_t comm_task, double computation)
679 char *mbox = bprintf("__mb_task_tx_overhead_%s", MSG_task_get_name(comm_task));
681 msg_task_t task = MSG_task_create("micro", computation, 0, NULL);
683 // XBT_INFO("req overhead");
684 msg_error_t ret = MSG_task_send(task, mbox);
685 xbt_assert(ret == MSG_OK);
690 /* alpha is (floating_operations / bytes).
692 * When actual migration traffic was 32 mbytes/s, we observed the CPU
693 * utilization of the main thread of the Qemu process was 10 %.
694 * alpha = 0.1 * C / (32 * 1024 * 1024)
695 * where the CPU capacity of the PM is C flops/s.
698 static void task_send_bounded_with_cpu_overhead(msg_task_t comm_task, char *mbox, double mig_speed, double alpha)
700 const double chunk_size = 1024 * 1024 * 10;
701 double remaining = MSG_task_get_data_size(comm_task);
703 start_overhead_process(comm_task);
706 while (remaining > 0) {
707 double data_size = chunk_size;
708 if (remaining < chunk_size)
709 data_size = remaining;
711 remaining -= data_size;
713 // XBT_INFO("remaining %f bytes", remaining);
716 double clock_sta = MSG_get_clock();
718 /* create a micro task */
720 char *mtask_name = bprintf("__micro_%s", MSG_task_get_name(comm_task));
721 msg_task_t mtask = MSG_task_create(mtask_name, 0, data_size, NULL);
723 request_overhead(comm_task, data_size * alpha);
725 msg_error_t ret = MSG_task_send(mtask, mbox);
726 xbt_assert(ret == MSG_OK);
728 xbt_free(mtask_name);
733 /* In the real world, sending data involves small CPU computation. */
734 char *mtask_name = bprintf("__micro_%s", MSG_task_get_name(comm_task));
735 msg_task_t mtask = MSG_task_create(mtask_name, data_size * alpha, data_size, NULL);
736 MSG_task_execute(mtask);
737 MSG_task_destroy(mtask);
738 xbt_free(mtask_name);
744 double clock_end = MSG_get_clock();
749 * (max bandwidth) > data_size / ((elapsed time) + time_to_sleep)
752 * time_to_sleep > data_size / (max bandwidth) - (elapsed time)
754 * If time_to_sleep is smaller than zero, the elapsed time was too big. We
755 * do not need a micro sleep.
757 double time_to_sleep = data_size / mig_speed - (clock_end - clock_sta);
758 if (time_to_sleep > 0)
759 MSG_process_sleep(time_to_sleep);
762 //XBT_INFO("duration %f", clock_end - clock_sta);
763 //XBT_INFO("time_to_sleep %f", time_to_sleep);
767 // XBT_INFO("%s", MSG_task_get_name(comm_task));
768 shutdown_overhead_process(comm_task);
774 static void make_cpu_overhead_of_data_transfer(msg_task_t comm_task, double init_comm_size)
776 double prev_remaining = init_comm_size;
779 double remaining = MSG_task_get_remaining_communication(comm_task);
783 double sent = prev_remaining - remaining;
784 double comp_size = sent * overhead;
787 char *comp_task_name = bprintf("__sender_overhead%s", MSG_task_get_name(comm_task));
788 msg_task_t comp_task = MSG_task_create(comp_task_name, comp_size, 0, NULL);
789 MSG_task_execute(comp_task);
790 MSG_task_destroy(comp_task);
795 prev_remaining = remaining;
799 xbt_free(comp_task_name);
803 // #define USE_MICRO_TASK 1
806 // const double alpha = 0.1L * 1.0E8 / (32L * 1024 * 1024);
807 // const double alpha = 0.25L * 1.0E8 / (85L * 1024 * 1024);
808 // const double alpha = 0.20L * 1.0E8 / (85L * 1024 * 1024);
809 // const double alpha = 0.25L * 1.0E8 / (85L * 1024 * 1024);
810 // const double alpha = 0.32L * 1.0E8 / (24L * 1024 * 1024); // makes super good values for 32 mbytes/s
811 //const double alpha = 0.32L * 1.0E8 / (32L * 1024 * 1024);
812 // const double alpha = 0.56L * 1.0E8 / (80L * 1024 * 1024);
813 ////const double alpha = 0.20L * 1.0E8 / (80L * 1024 * 1024);
814 // const double alpha = 0.56L * 1.0E8 / (90L * 1024 * 1024);
815 // const double alpha = 0.66L * 1.0E8 / (90L * 1024 * 1024);
816 // const double alpha = 0.20L * 1.0E8 / (80L * 1024 * 1024);
818 /* CPU 22% when 80Mbyte/s */
819 const double alpha = 0.22L * 1.0E8 / (80L * 1024 * 1024);
823 static void send_migration_data(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm,
824 sg_size_t size, char *mbox, int stage, int stage2_round, double mig_speed, double xfer_cpu_overhead)
826 char *task_name = get_mig_task_name(vm, src_pm, dst_pm, stage);
827 msg_task_t task = MSG_task_create(task_name, 0, size, NULL);
831 double clock_sta = MSG_get_clock();
833 #ifdef USE_MICRO_TASK
835 task_send_bounded_with_cpu_overhead(task, mbox, mig_speed, xfer_cpu_overhead);
840 ret = MSG_task_send_bounded(task, mbox, mig_speed);
842 ret = MSG_task_send(task, mbox);
843 xbt_assert(ret == MSG_OK);
846 double clock_end = MSG_get_clock();
847 double duration = clock_end - clock_sta;
848 double actual_speed = size / duration;
849 #ifdef USE_MICRO_TASK
850 double cpu_utilization = size * xfer_cpu_overhead / duration / 1.0E8;
852 double cpu_utilization = 0;
859 XBT_DEBUG("mig-stage%d.%d: sent %llu duration %f actual_speed %f (target %f) cpu %f", stage, stage2_round, size, duration, actual_speed, mig_speed, cpu_utilization);}
861 XBT_DEBUG("mig-stage%d: sent %llu duration %f actual_speed %f (target %f) cpu %f", stage, size, duration, actual_speed, mig_speed, cpu_utilization);
868 #ifdef USE_MICRO_TASK
869 /* The name of a micro task starts with __micro, which does not match the
870 * special name that finalizes the receiver loop. Thus, we send the special task.
874 char *task_name = get_mig_task_name(vm_name, src_pm_name, dst_pm_name, stage);
875 msg_task_t task = MSG_task_create(task_name, 0, 0, NULL);
876 msg_error_t ret = MSG_task_send(task, mbox);
877 xbt_assert(ret == MSG_OK);
884 static double get_updated_size(double computed, double dp_rate, double dp_cap)
886 double updated_size = computed * dp_rate;
887 XBT_DEBUG("updated_size %f dp_rate %f", updated_size, dp_rate);
888 if (updated_size > dp_cap) {
889 // XBT_INFO("mig-stage2.%d: %f bytes updated, but cap it with the working set size %f", stage2_round, updated_size, dp_cap);
890 updated_size = dp_cap;
896 static double send_stage1(struct migration_session *ms,
897 sg_size_t ramsize, double mig_speed, double xfer_cpu_overhead, double dp_rate, double dp_cap, double dpt_cpu_overhead)
900 // const long chunksize = (sg_size_t)1024 * 1024 * 100;
901 const sg_size_t chunksize = (sg_size_t)1024 * 1024 * 100000;
902 sg_size_t remaining = ramsize;
903 double computed_total = 0;
905 while (remaining > 0) {
906 sg_size_t datasize = chunksize;
907 if (remaining < chunksize)
908 datasize = remaining;
910 remaining -= datasize;
912 send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, datasize, ms->mbox, 1, 0, mig_speed, xfer_cpu_overhead);
913 double computed = lookup_computed_flop_counts(ms->vm, 1, 0);
914 computed_total += computed;
917 // double updated_size = get_updated_size(computed, dp_rate, dp_cap);
919 // double overhead = dpt_cpu_overhead * updated_size;
920 // launch_deferred_exec_process(vm, overhead, 10000);
924 return computed_total;
929 static double get_threshold_value(double bandwidth, double max_downtime)
931 /* This value assumes the network link is 1Gbps. */
932 // double threshold = max_downtime * 125 * 1024 * 1024;
933 double threshold = max_downtime * bandwidth;
938 static int migration_tx_fun(int argc, char *argv[])
940 XBT_DEBUG("mig: tx_start");
942 struct migration_session *ms = MSG_process_get_data(MSG_process_self());
944 s_ws_params_t params;
945 simcall_host_get_params(ms->vm, ¶ms);
946 const sg_size_t ramsize = params.ramsize;
947 const sg_size_t devsize = params.devsize;
948 const int skip_stage1 = params.skip_stage1;
949 const int skip_stage2 = params.skip_stage2;
950 const double dp_rate = params.dp_rate;
951 const double dp_cap = params.dp_cap;
952 const double mig_speed = params.mig_speed;
953 const double xfer_cpu_overhead = params.xfer_cpu_overhead;
954 const double dpt_cpu_overhead = params.dpt_cpu_overhead;
956 double remaining_size = ramsize + devsize;
958 double max_downtime = params.max_downtime;
959 if (max_downtime == 0) {
960 XBT_WARN("use the default max_downtime value 30ms");
964 double threshold = 0.00001; /* TODO: cleanup */
966 /* setting up parameters has done */
970 XBT_WARN("migrate a VM, but ramsize is zero");
973 XBT_INFO("mig-stage1: remaining_size %f", remaining_size);
975 /* Stage1: send all memory pages to the destination. */
976 start_dirty_page_tracking(ms->vm);
978 double computed_during_stage1 = 0;
980 // send_migration_data(vm_name, src_pm_name, dst_pm_name, ramsize, mbox, 1, 0, mig_speed, xfer_cpu_overhead);
982 /* send ramsize, but split it */
983 double clock_prev_send = MSG_get_clock();
985 computed_during_stage1 = send_stage1(ms, ramsize, mig_speed, xfer_cpu_overhead, dp_rate, dp_cap, dpt_cpu_overhead);
986 remaining_size -= ramsize;
988 double clock_post_send = MSG_get_clock();
989 double bandwidth = ramsize / (clock_post_send - clock_prev_send);
990 threshold = get_threshold_value(bandwidth, max_downtime);
991 XBT_INFO("actual bandwidth %f (MB/s), threshold %f", bandwidth / 1024 / 1024, threshold);
995 /* Stage2: send update pages iteratively until the size of remaining states
996 * becomes smaller than the threshold value. */
999 if (max_downtime == 0) {
1000 XBT_WARN("no max_downtime parameter, skip stage2");
1005 int stage2_round = 0;
1008 double updated_size = 0;
1009 if (stage2_round == 0) {
1010 /* just after stage1, nothing has been updated. But, we have to send the data updated during stage1 */
1011 updated_size = get_updated_size(computed_during_stage1, dp_rate, dp_cap);
1013 double computed = lookup_computed_flop_counts(ms->vm, 2, stage2_round);
1014 updated_size = get_updated_size(computed, dp_rate, dp_cap);
1017 XBT_INFO("mig-stage 2:%d updated_size %f computed_during_stage1 %f dp_rate %f dp_cap %f",
1018 stage2_round, updated_size, computed_during_stage1, dp_rate, dp_cap);
1021 // if (stage2_round != 0) {
1022 // /* during stage1, we have already created overhead tasks */
1023 // double overhead = dpt_cpu_overhead * updated_size;
1024 // XBT_DEBUG("updated %f overhead %f", updated_size, overhead);
1025 // launch_deferred_exec_process(vm, overhead, 10000);
1030 remaining_size += updated_size;
1032 XBT_INFO("mig-stage2.%d: remaining_size %f (%s threshold %f)", stage2_round,
1033 remaining_size, (remaining_size < threshold) ? "<" : ">", threshold);
1035 if (remaining_size < threshold)
1039 double clock_prev_send = MSG_get_clock();
1041 send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, updated_size, ms->mbox, 2, stage2_round, mig_speed, xfer_cpu_overhead);
1043 double clock_post_send = MSG_get_clock();
1045 double bandwidth = updated_size / (clock_post_send - clock_prev_send);
1046 threshold = get_threshold_value(bandwidth, max_downtime);
1047 XBT_INFO("actual bandwidth %f, threshold %f", bandwidth / 1024 / 1024, threshold);
1050 remaining_size -= updated_size;
1056 /* Stage3: stop the VM and copy the rest of states. */
1057 XBT_INFO("mig-stage3: remaining_size %f", remaining_size);
1058 simcall_vm_suspend(ms->vm);
1059 stop_dirty_page_tracking(ms->vm);
1061 send_migration_data(ms->vm, ms->src_pm, ms->dst_pm, remaining_size, ms->mbox, 3, 0, mig_speed, xfer_cpu_overhead);
1064 XBT_DEBUG("mig: tx_done");
1071 static void do_migration(msg_vm_t vm, msg_host_t src_pm, msg_host_t dst_pm)
1073 struct migration_session *ms = xbt_new(struct migration_session, 1);
1075 ms->src_pm = src_pm;
1076 ms->dst_pm = dst_pm;
1077 ms->mbox_ctl = get_mig_mbox_ctl(vm, src_pm, dst_pm);
1078 ms->mbox = get_mig_mbox_src_dst(vm, src_pm, dst_pm);
1080 char *pr_rx_name = get_mig_process_rx_name(vm, src_pm, dst_pm);
1081 char *pr_tx_name = get_mig_process_tx_name(vm, src_pm, dst_pm);
1083 // MSG_process_create(pr_rx_name, migration_rx_fun, ms, dst_pm);
1084 // MSG_process_create(pr_tx_name, migration_tx_fun, ms, src_pm);
1087 char **argv = xbt_new(char *, 2);
1088 argv[0] = pr_rx_name;
1090 MSG_process_create_with_arguments(pr_rx_name, migration_rx_fun, ms, dst_pm, 1, argv);
1093 char **argv = xbt_new(char *, 2);
1094 argv[0] = pr_tx_name;
1096 MSG_process_create_with_arguments(pr_tx_name, migration_tx_fun, ms, src_pm, 1, argv);
1103 /* wait until the migration have finished */
1105 msg_task_t task = NULL;
1106 msg_error_t ret = MSG_task_recv(&task, ms->mbox_ctl);
1108 xbt_assert(ret == MSG_OK);
1110 char *expected_task_name = get_mig_task_name(vm, src_pm, dst_pm, 4);
1111 xbt_assert(strcmp(task->name, expected_task_name) == 0);
1112 xbt_free(expected_task_name);
1113 MSG_task_destroy(task);
1116 xbt_free(ms->mbox_ctl);
1122 /** @brief Migrate the VM to the given host.
1125 * FIXME: No migration cost occurs. If you want to simulate this too, you want to use a
1126 * MSG_task_send() before or after, depending on whether you want to do cold or hot
1129 void MSG_vm_migrate(msg_vm_t vm, msg_host_t new_pm)
1132 * - One approach is ...
1133 * We first create a new VM (i.e., destination VM) on the destination
1134 * physical host. The destination VM will receive the state of the source
1135 * VM over network. We will finally destroy the source VM.
1136 * - This behavior is similar to the way of migration in the real world.
1137 * Even before a migration is completed, we will see a destination VM,
1138 * consuming resources.
1139 * - We have to relocate all processes. The existing process migraion code
1140 * will work for this?
1141 * - The name of the VM is a somewhat unique ID in the code. It is tricky
1142 * for the destination VM?
1144 * - Another one is ...
1145 * We update the information of the given VM to place it to the destination
1148 * The second one would be easier.
1152 msg_host_t old_pm = simcall_vm_get_pm(vm);
1154 if (!MSG_vm_is_running(vm))
1155 THROWF(vm_error, 0, "VM(%s) is not running", sg_host_name(vm));
1157 if (MSG_vm_is_migrating(vm))
1158 THROWF(vm_error, 0, "VM(%s) is already migrating", sg_host_name(vm));
1160 msg_host_priv_t priv = msg_host_resource_priv(vm);
1161 priv->is_migrating = 1;
1163 do_migration(vm, old_pm, new_pm);
1165 priv->is_migrating = 0;
1167 XBT_DEBUG("VM(%s) moved from PM(%s) to PM(%s)", vm->key, old_pm->key, new_pm->key);
1170 TRACE_msg_vm_change_host(vm, old_pm, new_pm);
1175 /** @brief Immediately suspend the execution of all processes within the given VM.
1178 * This function stops the execution of the VM. All the processes on this VM
1179 * will pause. The state of the VM is preserved. We can later resume it again.
1181 * No suspension cost occurs.
1183 void MSG_vm_suspend(msg_vm_t vm)
1185 if (MSG_vm_is_migrating(vm))
1186 THROWF(vm_error, 0, "VM(%s) is migrating", sg_host_name(vm));
1188 simcall_vm_suspend(vm);
1190 XBT_DEBUG("vm_suspend done");
1193 TRACE_msg_vm_suspend(vm);
1198 /** @brief Resume the execution of the VM. All processes on the VM run again.
1201 * No resume cost occurs.
1203 void MSG_vm_resume(msg_vm_t vm)
1205 simcall_vm_resume(vm);
1208 TRACE_msg_vm_resume(vm);
1213 /** @brief Immediately save the execution of all processes within the given VM.
1216 * This function stops the execution of the VM. All the processes on this VM
1217 * will pause. The state of the VM is preserved. We can later resume it again.
1219 * FIXME: No suspension cost occurs. If you want to simulate this too, you want to
1220 * use a \ref MSG_file_write() before or after, depending on the exact semantic
1221 * of VM save to you.
1223 void MSG_vm_save(msg_vm_t vm)
1225 if (MSG_vm_is_migrating(vm))
1226 THROWF(vm_error, 0, "VM(%s) is migrating", sg_host_name(vm));
1228 simcall_vm_save(vm);
1230 TRACE_msg_vm_save(vm);
1234 /** @brief Restore the execution of the VM. All processes on the VM run again.
1237 * FIXME: No restore cost occurs. If you want to simulate this too, you want to
1238 * use a \ref MSG_file_read() before or after, depending on the exact semantic
1239 * of VM restore to you.
1241 void MSG_vm_restore(msg_vm_t vm)
1243 simcall_vm_restore(vm);
1246 TRACE_msg_vm_restore(vm);
1251 /** @brief Get the physical host of a given VM.
1254 msg_host_t MSG_vm_get_pm(msg_vm_t vm)
1256 return simcall_vm_get_pm(vm);
1260 /** @brief Set a CPU bound for a given VM.
1264 * Note that in some cases MSG_task_set_bound() may not intuitively work for VMs.
1267 * On PM0, there are Task1 and VM0.
1268 * On VM0, there is Task2.
1269 * Now we bound 75% to Task1\@PM0 and bound 25% to Task2\@VM0.
1271 * Task1\@PM0 gets 50%.
1272 * Task2\@VM0 gets 25%.
1273 * This is NOT 75% for Task1\@PM0 and 25% for Task2\@VM0, respectively.
1275 * This is because a VM has the dummy CPU action in the PM layer. Putting a
1276 * task on the VM does not affect the bound of the dummy CPU action. The bound
1277 * of the dummy CPU action is unlimited.
1279 * There are some solutions for this problem. One option is to update the bound
1280 * of the dummy CPU action automatically. It should be the sum of all tasks on
1281 * the VM. But, this solution might be costly, because we have to scan all tasks
1282 * on the VM in share_resource() or we have to trap both the start and end of
1285 * The current solution is to use MSG_vm_set_bound(), which allows us to
1286 * directly set the bound of the dummy CPU action.
1290 * Note that bound == 0 means no bound (i.e., unlimited). But, if a host has
1291 * multiple CPU cores, the CPU share of a computation task (or a VM) never
1292 * exceeds the capacity of a CPU core.
1294 void MSG_vm_set_bound(msg_vm_t vm, double bound)
1296 return simcall_vm_set_bound(vm, bound);
1300 /** @brief Set the CPU affinity of a given VM.
1303 * This function changes the CPU affinity of a given VM. Usage is the same as
1304 * MSG_task_set_affinity(). See the MSG_task_set_affinity() for details.
1306 void MSG_vm_set_affinity(msg_vm_t vm, msg_host_t pm, unsigned long mask)
1308 msg_host_priv_t priv = msg_host_resource_priv(vm);
1311 xbt_dict_remove_ext(priv->affinity_mask_db, (char *) pm, sizeof(pm));
1313 xbt_dict_set_ext(priv->affinity_mask_db, (char *) pm, sizeof(pm), (void *) mask, NULL);
1315 msg_host_t pm_now = MSG_vm_get_pm(vm);
1317 XBT_INFO("set affinity(0x%04lx@%s) for %s", mask, MSG_host_get_name(pm), MSG_host_get_name(vm));
1318 simcall_vm_set_affinity(vm, pm, mask);
1320 XBT_INFO("set affinity(0x%04lx@%s) for %s (not active now)", mask, MSG_host_get_name(pm), MSG_host_get_name(vm));