1 /* Copyright (c) 2013-2023. The SimGrid Team. All rights reserved. */
3 /* This program is free software; you can redistribute it and/or modify it
4 * under the terms of the license (GNU LGPL) which comes with this package. */
6 #include <simgrid/Exception.hpp>
7 #include <simgrid/kernel/routing/NetPoint.hpp>
8 #include <simgrid/kernel/routing/NetZoneImpl.hpp>
9 #include <simgrid/s4u/Exec.hpp>
11 #include "src/kernel/EngineImpl.hpp"
12 #include "src/kernel/activity/ExecImpl.hpp"
13 #include "src/kernel/resource/VirtualMachineImpl.hpp"
14 #include "src/kernel/resource/models/cpu_cas01.hpp"
15 #include "src/kernel/resource/models/cpu_ti.hpp"
16 #include "src/simgrid/module.hpp"
17 #include "src/simgrid/sg_config.hpp"
21 XBT_LOG_NEW_DEFAULT_SUBCATEGORY(res_vm, ker_resource, "Virtual Machines, containing actors and mobile across hosts");
23 void simgrid_vm_model_init_HL13()
25 auto* cpu_pm_model = simgrid::s4u::Engine::get_instance()->get_netzone_root()->get_impl()->get_cpu_pm_model().get();
26 auto vm_model = std::make_shared<simgrid::kernel::resource::VMModel>("VM_HL13");
27 auto* engine = simgrid::kernel::EngineImpl::get_instance();
29 engine->add_model(vm_model, {cpu_pm_model});
30 std::shared_ptr<simgrid::kernel::resource::CpuModel> cpu_model_vm;
32 if (simgrid::config::get_value<std::string>("cpu/optim") == "TI") {
33 cpu_model_vm = std::make_shared<simgrid::kernel::resource::CpuTiModel>("VmCpu_TI");
35 cpu_model_vm = std::make_shared<simgrid::kernel::resource::CpuCas01Model>("VmCpu_Cas01");
37 engine->add_model(cpu_model_vm, {cpu_pm_model, vm_model.get()});
38 engine->get_netzone_root()->set_cpu_vm_model(cpu_model_vm);
42 template class xbt::Extendable<kernel::resource::VirtualMachineImpl>;
44 namespace kernel::resource {
50 std::deque<s4u::VirtualMachine*> VirtualMachineImpl::allVms_;
52 /* In the real world, processes on the guest operating system will be somewhat degraded due to virtualization overhead.
53 * The total CPU share these processes get is smaller than that of the VM process gets on a host operating system.
54 * FIXME: add a configuration flag for this
56 const double virt_overhead = 1; // 0.95
58 static void host_onoff(s4u::Host const& host)
60 if (not host.is_on()) { // just turned off.
61 std::vector<s4u::VirtualMachine*> trash;
62 /* Find all VMs living on that host */
63 for (auto* vm : VirtualMachineImpl::allVms_)
64 if (vm->get_pm() == &host)
66 for (auto* vm : trash)
71 static void add_active_exec(s4u::Exec const& task)
73 const s4u::VirtualMachine* vm = dynamic_cast<s4u::VirtualMachine*>(task.get_host());
75 VirtualMachineImpl* vm_impl = vm->get_vm_impl();
76 for (int i = 1; i <= task.get_thread_count(); i++)
77 vm_impl->add_active_exec();
78 vm_impl->update_action_weight();
82 static void remove_active_exec(s4u::Exec const& exec)
84 if (not exec.is_assigned())
86 const s4u::VirtualMachine* vm = dynamic_cast<s4u::VirtualMachine*>(exec.get_host());
88 VirtualMachineImpl* vm_impl = vm->get_vm_impl();
89 for (int i = 1; i <= exec.get_thread_count(); i++)
90 vm_impl->remove_active_exec();
91 vm_impl->update_action_weight();
95 static s4u::VirtualMachine* get_vm_from_activity(s4u::Activity const& act)
97 auto* exec = dynamic_cast<kernel::activity::ExecImpl const*>(act.get_impl());
98 return exec != nullptr ? dynamic_cast<s4u::VirtualMachine*>(exec->get_host()) : nullptr;
101 static void add_active_activity(s4u::Activity const& act)
103 const s4u::VirtualMachine* vm = get_vm_from_activity(act);
105 VirtualMachineImpl* vm_impl = vm->get_vm_impl();
106 vm_impl->add_active_exec();
107 vm_impl->update_action_weight();
111 static void remove_active_activity(s4u::Activity const& act)
113 const s4u::VirtualMachine* vm = get_vm_from_activity(act);
115 VirtualMachineImpl* vm_impl = vm->get_vm_impl();
116 vm_impl->remove_active_exec();
117 vm_impl->update_action_weight();
121 VMModel::VMModel(const std::string& name) : HostModel(name)
123 s4u::Host::on_onoff_cb(host_onoff);
124 s4u::Exec::on_start_cb(add_active_exec);
125 s4u::Exec::on_completion_cb(remove_active_exec);
126 s4u::Exec::on_resume_cb(add_active_activity);
127 s4u::Exec::on_suspend_cb(remove_active_activity);
130 double VMModel::next_occurring_event(double now)
132 /* TODO: update action's cost with the total cost of processes on the VM. */
134 /* 1. Now we know how many resource should be assigned to each virtual
135 * machine. We update constraints of the virtual machine layer.
137 * If we have two virtual machine (VM1 and VM2) on a physical machine (PM1).
138 * X1 + X2 = C (Equation 1)
140 * the resource share of VM1: X1
141 * the resource share of VM2: X2
142 * the capacity of PM1: C
144 * Then, if we have two process (P1 and P2) on VM1.
145 * X1_1 + X1_2 = X1 (Equation 2)
147 * the resource share of P1: X1_1
148 * the resource share of P2: X1_2
149 * the capacity of VM1: X1
151 * Equation 1 was solved in the physical machine layer.
152 * Equation 2 is solved in the virtual machine layer (here).
153 * X1 must be passed to the virtual machine layer as a constraint value.
156 /* iterate for all virtual machines */
157 for (auto const* ws_vm : VirtualMachineImpl::allVms_) {
158 if (ws_vm->get_state() == s4u::VirtualMachine::State::SUSPENDED) // Ignore suspended VMs
161 const kernel::resource::CpuImpl* cpu = ws_vm->get_cpu();
163 // solved_value below is X1 in comment above: what this VM got in the sharing on the PM
164 double solved_value = ws_vm->get_vm_impl()->get_action()->get_rate();
165 XBT_DEBUG("assign %f to vm %s @ pm %s", solved_value, ws_vm->get_cname(), ws_vm->get_pm()->get_cname());
167 lmm::System* vcpu_system = cpu->get_model()->get_maxmin_system();
168 vcpu_system->update_constraint_bound(cpu->get_constraint(), virt_overhead * solved_value);
170 /* actual next occurring event is determined by VM CPU model at EngineImpl::solve */
174 Action* VMModel::execute_thread(const s4u::Host* host, double flops_amount, int thread_count)
176 auto cpu = host->get_cpu();
177 return cpu->execution_start(thread_count * flops_amount, thread_count, -1);
184 VirtualMachineImpl::VirtualMachineImpl(const std::string& name, s4u::VirtualMachine* piface,
185 simgrid::s4u::Host* host_PM, int core_amount, size_t ramsize)
186 : VirtualMachineImpl(name, host_PM, core_amount, ramsize)
191 VirtualMachineImpl::VirtualMachineImpl(const std::string& name, simgrid::s4u::Host* host_PM, int core_amount,
193 : HostImpl(name), physical_host_(host_PM), core_amount_(core_amount), ramsize_(ramsize)
195 /* We create cpu_action corresponding to a VM process on the host operating system. */
196 /* TODO: we have to periodically input GUESTOS_NOISE to the system? how ?
197 * The value for GUESTOS_NOISE corresponds to the cost of the global action associated to the VM. It corresponds to
198 * the cost of a VM running no tasks.
200 action_ = physical_host_->get_cpu()->execution_start(0, core_amount_, 0);
202 // It's empty for now, so it should not request resources in the PM
203 update_action_weight();
204 XBT_VERB("Create VM(%s)@PM(%s)", name.c_str(), physical_host_->get_cname());
207 void VirtualMachineImpl::set_piface(s4u::VirtualMachine* piface)
209 xbt_assert(not piface_, "Pointer to interface already configured for this VM (%s)", get_cname());
211 /* Register this VM to the list of all VMs */
212 allVms_.push_back(piface);
215 /** @brief A physical host does not disappear in the current SimGrid code, but a VM may disappear during a simulation */
216 void VirtualMachineImpl::vm_destroy()
218 /* I was already removed from the allVms set if the VM was destroyed cleanly */
219 if (auto iter = find(allVms_.begin(), allVms_.end(), piface_); iter != allVms_.end())
222 /* Free the cpu_action of the VM. */
223 XBT_ATTRIB_UNUSED bool ret = action_->unref();
224 xbt_assert(ret, "Bug: some resource still remains");
226 // VM uses the host's netpoint, clean but don't destroy it
227 get_iface()->set_netpoint(nullptr);
228 // Take a temporary copy to delete iface safely after impl is destroy'ed
229 const auto* iface = get_iface();
230 // calls the HostImpl() destroy, it'll delete the impl object
236 void VirtualMachineImpl::start()
238 s4u::VirtualMachine::on_start(*get_iface());
239 get_iface()->on_this_start(*get_iface());
240 s4u::VmHostExt::ensureVmExtInstalled();
242 if (physical_host_->extension<s4u::VmHostExt>() == nullptr)
243 physical_host_->extension_set(new s4u::VmHostExt());
245 if (size_t pm_ramsize = physical_host_->extension<s4u::VmHostExt>()->ramsize;
247 not physical_host_->extension<s4u::VmHostExt>()->overcommit) { /* Need to verify that we don't overcommit */
248 /* Retrieve the memory occupied by the VMs on that host. Yep, we have to traverse all VMs of all hosts for that */
249 size_t total_ramsize_of_vms = 0;
250 for (auto const* ws_vm : allVms_)
251 if (physical_host_ == ws_vm->get_pm())
252 total_ramsize_of_vms += ws_vm->get_ramsize();
254 if (total_ramsize_of_vms + get_ramsize() > pm_ramsize) {
255 XBT_WARN("cannot start %s@%s due to memory shortage: get_ramsize() %zu, free %zu, pm_ramsize %zu (bytes).",
256 get_cname(), physical_host_->get_cname(), get_ramsize(), pm_ramsize - total_ramsize_of_vms, pm_ramsize);
257 throw VmFailureException(XBT_THROW_POINT,
258 xbt::string_printf("Memory shortage on host '%s', VM '%s' cannot be started",
259 physical_host_->get_cname(), get_cname()));
262 vm_state_ = s4u::VirtualMachine::State::RUNNING;
264 s4u::VirtualMachine::on_started(*get_iface());
265 get_iface()->on_this_started(*get_iface());
268 void VirtualMachineImpl::suspend(const actor::ActorImpl* issuer)
270 s4u::VirtualMachine::on_suspend(*get_iface());
271 get_iface()->on_this_suspend(*get_iface());
273 if (vm_state_ != s4u::VirtualMachine::State::RUNNING)
274 throw VmFailureException(XBT_THROW_POINT,
275 xbt::string_printf("Cannot suspend VM %s: it is not running.", piface_->get_cname()));
276 if (issuer->get_host() == piface_)
277 throw VmFailureException(XBT_THROW_POINT, xbt::string_printf("Actor %s cannot suspend the VM %s in which it runs",
278 issuer->get_cname(), piface_->get_cname()));
280 XBT_DEBUG("suspend VM(%s), where %zu actors exist", piface_->get_cname(), get_actor_count());
284 foreach_actor([](auto& actor) {
285 XBT_DEBUG("suspend %s", actor.get_cname());
289 XBT_DEBUG("suspend all actors on the VM done done");
291 vm_state_ = s4u::VirtualMachine::State::SUSPENDED;
294 void VirtualMachineImpl::resume()
296 if (vm_state_ != s4u::VirtualMachine::State::SUSPENDED)
297 throw VmFailureException(XBT_THROW_POINT,
298 xbt::string_printf("Cannot resume VM %s: it was not suspended", piface_->get_cname()));
300 XBT_DEBUG("Resume VM %s, containing %zu actors.", piface_->get_cname(), get_actor_count());
304 foreach_actor([](auto& actor) {
305 XBT_DEBUG("resume %s", actor.get_cname());
309 vm_state_ = s4u::VirtualMachine::State::RUNNING;
310 s4u::VirtualMachine::on_resume(*get_iface());
311 get_iface()->on_this_resume(*get_iface());
314 /** @brief Power off a VM.
316 * All hosted processes will be killed, but the VM state is preserved on memory.
317 * It can later be restarted.
319 * @param issuer the actor requesting the shutdown
321 void VirtualMachineImpl::shutdown(actor::ActorImpl* issuer)
323 if (vm_state_ != s4u::VirtualMachine::State::RUNNING)
324 XBT_VERB("Shutting down the VM %s even if it's not running but in state %s", piface_->get_cname(),
325 s4u::VirtualMachine::to_c_str(get_state()));
327 XBT_DEBUG("shutdown VM %s, that contains %zu actors", piface_->get_cname(), get_actor_count());
329 foreach_actor([issuer](auto& actor) {
330 XBT_DEBUG("kill %s@%s on behalf of %s which shutdown that VM.", actor.get_cname(), actor.get_host()->get_cname(),
331 issuer->get_cname());
332 issuer->kill(&actor);
335 set_state(s4u::VirtualMachine::State::DESTROYED);
337 s4u::VirtualMachine::on_shutdown(*get_iface());
338 get_iface()->on_this_shutdown(*get_iface());
341 /** @brief Change the physical host on which the given VM is running
343 * This is an instantaneous migration.
345 void VirtualMachineImpl::set_physical_host(s4u::Host* destination)
347 std::string vm_name = piface_->get_name();
348 std::string pm_name_src = physical_host_->get_name();
349 std::string pm_name_dst = destination->get_name();
351 /* update net_elm with that of the destination physical host */
352 piface_->set_netpoint(destination->get_netpoint());
353 physical_host_->get_impl()->move_vm(this, destination->get_impl());
355 /* Adapt the speed, pstate and other physical characteristics to the one of our new physical CPU */
356 piface_->get_cpu()->reset_vcpu(destination->get_cpu());
358 physical_host_ = destination;
360 /* Update vcpu's action for the new pm */
361 /* create a cpu action bound to the pm model at the destination. */
362 CpuAction* new_cpu_action = destination->get_cpu()->execution_start(0, this->core_amount_);
364 if (action_->get_remains_no_update() > 0)
365 XBT_CRITICAL("FIXME: need copy the state(?), %f", action_->get_remains_no_update());
367 /* keep the bound value of the cpu action of the VM. */
368 if (double old_bound = action_->get_bound(); old_bound > 0) {
369 XBT_DEBUG("migrate VM(%s): set bound (%f) at %s", vm_name.c_str(), old_bound, pm_name_dst.c_str());
370 new_cpu_action->set_bound(old_bound);
373 XBT_ATTRIB_UNUSED bool ret = action_->unref();
374 xbt_assert(ret, "Bug: some resource still remains");
376 action_ = new_cpu_action;
378 XBT_DEBUG("migrate VM(%s): change PM (%s to %s)", vm_name.c_str(), pm_name_src.c_str(), pm_name_dst.c_str());
381 void VirtualMachineImpl::set_bound(double bound)
384 action_->set_user_bound(user_bound_);
385 update_action_weight();
388 void VirtualMachineImpl::update_action_weight()
390 /* The impact of the VM over its PM is the min between its vCPU amount and the amount of tasks it contains */
391 int impact = std::min(active_execs_, get_core_amount());
393 XBT_DEBUG("set the weight of the dummy CPU action of VM%p on PM to %d (#tasks: %u)", this, impact, active_execs_);
396 action_->set_sharing_penalty(1. / impact);
398 action_->set_sharing_penalty(0.);
400 action_->set_bound(std::min(impact * physical_host_->get_speed(), user_bound_));
403 void VirtualMachineImpl::start_migration()
405 is_migrating_ = true;
406 s4u::VirtualMachine::on_migration_start(*get_iface());
407 get_iface()->on_this_migration_start(*get_iface());
410 void VirtualMachineImpl::end_migration()
412 is_migrating_ = false;
413 s4u::VirtualMachine::on_migration_end(*get_iface());
414 get_iface()->on_this_migration_end(*get_iface());
417 void VirtualMachineImpl::seal()
420 s4u::VirtualMachine::on_vm_creation(*get_iface());
423 } // namespace kernel::resource
424 } // namespace simgrid