X-Git-Url: http://info.iut-bm.univ-fcomte.fr/pub/gitweb/simgrid.git/blobdiff_plain/75e8b2ae65722d849ef899616de7091286ec91f5..e2800b5e3f776e413749ce2d3e34fc4766f74a4a:/src/s4u/s4u_Host.cpp diff --git a/src/s4u/s4u_Host.cpp b/src/s4u/s4u_Host.cpp index 3f737173cc..8a70f85374 100644 --- a/src/s4u/s4u_Host.cpp +++ b/src/s4u/s4u_Host.cpp @@ -1,10 +1,12 @@ -/* Copyright (c) 2006-2019. The SimGrid Team. All rights reserved. */ +/* Copyright (c) 2006-2021. The SimGrid Team. All rights reserved. */ /* This program is free software; you can redistribute it and/or modify it * under the terms of the license (GNU LGPL) which comes with this package. */ +#include "simgrid/host.h" #include "simgrid/kernel/routing/NetPoint.hpp" #include "simgrid/s4u/Actor.hpp" +#include "simgrid/s4u/Comm.hpp" #include "simgrid/s4u/Engine.hpp" #include "simgrid/s4u/Exec.hpp" #include "simgrid/s4u/VirtualMachine.hpp" @@ -18,9 +20,8 @@ XBT_LOG_NEW_DEFAULT_SUBCATEGORY(s4u_host, s4u, "Logging specific to the S4U host XBT_LOG_EXTERNAL_CATEGORY(surf_route); namespace simgrid { -namespace xbt { -template class Extendable; -} // namespace xbt + +template class xbt::Extendable; namespace s4u { @@ -51,7 +52,7 @@ Host::~Host() * * Don't delete directly a host, call h->destroy() instead. * - * This is cumbersome but this is the simplest solution to ensure that the onDestruction() callback receives a valid + * This is cumbersome but this is the simplest solution to ensure that the on_destruction() callback receives a valid * object (because of the destructor order in a class hierarchy). */ void Host::destroy() @@ -96,7 +97,7 @@ void Host::turn_on() void Host::turn_off() { if (is_on()) { - kernel::actor::ActorImpl* self = kernel::actor::ActorImpl::self(); + const kernel::actor::ActorImpl* self = kernel::actor::ActorImpl::self(); kernel::actor::simcall([this, self] { for (VirtualMachine* const& vm : vm::VirtualMachineImpl::allVms_) if (vm->get_pm() == this) { @@ -153,7 +154,7 @@ void Host::route_to(const Host* dest, std::vector& links, double* latency { std::vector linkImpls; this->route_to(dest, linkImpls, latency); - for (kernel::resource::LinkImpl* const& l : linkImpls) + for (auto* l : linkImpls) links.push_back(l->get_iface()); } @@ -164,8 +165,8 @@ void Host::route_to(const Host* dest, std::vector& if (XBT_LOG_ISENABLED(surf_route, xbt_log_priority_debug)) { XBT_CDEBUG(surf_route, "Route from '%s' to '%s' (latency: %f):", get_cname(), dest->get_cname(), (latency == nullptr ? -1 : *latency)); - for (auto const& link : links) - XBT_CDEBUG(surf_route, "Link %s", link->get_cname()); + for (auto const* link : links) + XBT_CDEBUG(surf_route, " Link '%s'", link->get_cname()); } } @@ -175,12 +176,14 @@ NetZone* Host::get_englobing_zone() return pimpl_netpoint_->get_englobing_zone()->get_iface(); } -void Host::send_to(Host* dest, double byte_amount) +void Host::sendto(Host* dest, double byte_amount) +{ + sendto_async(dest, byte_amount)->wait(); +} + +CommPtr Host::sendto_async(Host* dest, double byte_amount) { - std::vector m_host_list = {this, dest}; - std::vector flops_amount = {0, 0}; - std::vector bytes_amount = {0, byte_amount, 0, 0}; - this_actor::parallel_execute(m_host_list, flops_amount, bytes_amount); + return Comm::sendto_async(this, dest, byte_amount); } /** Get the properties assigned to a host */ @@ -200,18 +203,18 @@ void Host::set_property(const std::string& key, const std::string& value) kernel::actor::simcall([this, &key, &value] { this->pimpl_->set_property(key, value); }); } -void Host::set_properties(const std::map& properties) +void Host::set_properties(const std::unordered_map& properties) { kernel::actor::simcall([this, &properties] { this->pimpl_->set_properties(properties); }); } -/** Specify a profile turning the host on and off according to a exhaustive list or a stochastic law. +/** Specify a profile turning the host on and off according to an exhaustive list or a stochastic law. * The profile must contain boolean values. */ void Host::set_state_profile(kernel::profile::Profile* p) { return kernel::actor::simcall([this, p] { pimpl_cpu->set_state_profile(p); }); } -/** Specify a profile modeling the external load according to a exhaustive list or a stochastic law. +/** Specify a profile modeling the external load according to an exhaustive list or a stochastic law. * * Each event of the profile represent a peak speed change that is due to external load. The values are given as a rate * of the initial value. This means that the actual value is obtained by multiplying the initial value (the peek speed @@ -228,44 +231,19 @@ double Host::get_pstate_speed(int pstate_index) const return this->pimpl_cpu->get_pstate_peak_speed(pstate_index); } -/** @brief Get the peak computing speed in flops/s at the current pstate, NOT taking the external load into account. - * - * The amount of flops per second available for computing depends on several things: - * - The current pstate determines the maximal peak computing speed (use @ref get_pstate_speed() to retrieve the - * computing speed you would get at another pstate) - * - If you declared an external load (with @ref simgrid::surf::Cpu::set_speed_profile()), you must multiply the - * result of get_speed() by get_available_speed() to retrieve what a new computation would get. - * - * The remaining speed is then shared between the executions located on this host. - * You can retrieve the amount of tasks currently running on this host with @ref get_load(). - * - * The host may have multiple cores, and your executions may be able to use more than a single core. - * - * Finally, executions of priority 2 get twice the amount of flops than executions of priority 1. - */ double Host::get_speed() const { return this->pimpl_cpu->get_speed(1.0); } -/** @brief Returns the current computation load (in flops per second) - * - * The external load (coming from an availability trace) is not taken in account. - * You may also be interested in the load plugin. - */ double Host::get_load() const { return this->pimpl_cpu->get_load(); } -/** @brief Get the available speed ratio, between 0 and 1. - * - * This accounts for external load (see @ref simgrid::surf::Cpu::set_speed_profile()). - */ double Host::get_available_speed() const { return this->pimpl_cpu->get_speed_ratio(); } -/** @brief Returns the number of core of the processor. */ int Host::get_core_count() const { return this->pimpl_cpu->get_core_count(); @@ -287,7 +265,7 @@ std::vector Host::get_disks() const return kernel::actor::simcall([this] { return this->pimpl_->get_disks(); }); } -void Host::add_disk(Disk* disk) +void Host::add_disk(const Disk* disk) { kernel::actor::simcall([this, disk] { this->pimpl_->add_disk(disk); }); } @@ -308,26 +286,24 @@ std::vector Host::get_attached_storages() const std::unordered_map const& Host::get_mounted_storages() { - if (mounts_ == nullptr) { - mounts_ = new std::unordered_map(); - for (auto const& m : this->pimpl_->storage_) { - mounts_->insert({m.first, m.second->get_iface()}); - } - } + kernel::actor::simcall([this] { + if (mounts_ == nullptr) + mounts_ = pimpl_->get_mounted_storages(); + }); return *mounts_; } -ExecPtr Host::exec_async(double flops) +ExecPtr Host::exec_async(double flops) const { return this_actor::exec_init(flops); } -void Host::execute(double flops) +void Host::execute(double flops) const { execute(flops, 1.0 /* priority */); } -void Host::execute(double flops, double priority) +void Host::execute(double flops, double priority) const { this_actor::exec_init(flops)->set_priority(1 / priority)->start()->wait(); } @@ -340,20 +316,12 @@ size_t sg_host_count() { return simgrid::s4u::Engine::get_instance()->get_host_count(); } -/** @brief Returns the host list - * - * Uses sg_host_count() to know the array size. - * - * @return an array of @ref sg_host_t containing all the hosts in the platform. - * @remark The host order in this array is generally different from the - * creation/declaration order in the XML platform (we use a hash table - * internally). - * @see sg_host_count() - */ sg_host_t* sg_host_list() { - xbt_assert(sg_host_count() > 0, "There is no host!"); - std::vector hosts = simgrid::s4u::Engine::get_instance()->get_all_hosts(); + const simgrid::s4u::Engine* e = simgrid::s4u::Engine::get_instance(); + size_t host_count = e->get_host_count(); + xbt_assert(host_count > 0, "There is no host!"); + std::vector hosts = e->get_all_hosts(); sg_host_t* res = xbt_new(sg_host_t, hosts.size()); memcpy(res, hosts.data(), sizeof(sg_host_t) * hosts.size()); @@ -381,7 +349,7 @@ sg_host_t sg_host_by_name(const char* name) return simgrid::s4u::Host::by_name_or_null(name); } -xbt_dynar_t sg_hosts_as_dynar() +xbt_dynar_t sg_hosts_as_dynar() // XBT_ATTRIB_DEPRECATED_v330 { std::vector list = simgrid::s4u::Engine::get_instance()->get_all_hosts(); @@ -399,29 +367,46 @@ xbt_dynar_t sg_hosts_as_dynar() // ========= Layering madness ==============* // ========== User data Layer ========== -void* sg_host_data(const_sg_host_t host) +void* sg_host_get_data(const_sg_host_t host) { return host->get_data(); } -void sg_host_data_set(sg_host_t host, void* userdata) +void sg_host_set_data(sg_host_t host, void* userdata) { host->set_data(userdata); } -void* sg_host_user(sg_host_t host) // deprecated +void* sg_host_data(const_sg_host_t host) // XBT_ATTRIB_DEPRECATED_v330 +{ + return sg_host_get_data(host); +} +void sg_host_data_set(sg_host_t host, void* userdata) // XBT_ATTRIB_DEPRECATED_v330 +{ + sg_host_set_data(host, userdata); +} +void* sg_host_user(sg_host_t host) // XBT_ATTRIB_DEPRECATED_v328 { return host->get_data(); } -void sg_host_user_set(sg_host_t host, void* userdata) // deprecated +void sg_host_user_set(sg_host_t host, void* userdata) // XBT_ATTRIB_DEPRECATED_v328 { host->set_data(userdata); } -void sg_host_user_destroy(sg_host_t host) // deprecated +void sg_host_user_destroy(sg_host_t host) // XBT_ATTRIB_DEPRECATED_v328 { host->set_data(nullptr); } // ========= storage related functions ============ -xbt_dict_t sg_host_get_mounted_storage_list(sg_host_t host) +void sg_host_get_disks(const_sg_host_t host, unsigned int* disk_count, sg_disk_t** disks) +{ + std::vector list = host->get_disks(); + *disk_count = list.size(); + *disks = static_cast(xbt_malloc(sizeof(sg_disk_t) * (*disk_count))); + for (size_t i = 0; i < *disk_count; i++) + (*disks)[i] = list[i]; +} + +xbt_dict_t sg_host_get_mounted_storage_list(sg_host_t host) // XBT_ATTRIB_DEPRECATED_v330 { xbt_assert((host != nullptr), "Invalid parameters"); xbt_dict_t res = xbt_dict_new_homogeneous(nullptr); @@ -446,11 +431,16 @@ xbt_dynar_t sg_host_get_attached_storage_list(const_sg_host_t host) // =========== user-level functions =============== // ================================================ /** @brief Returns the total speed of a host */ -double sg_host_speed(const_sg_host_t host) +double sg_host_get_speed(const_sg_host_t host) { return host->get_speed(); } +double sg_host_speed(const_sg_host_t host) // XBT_ATTRIB_DEPRECATED_v330 +{ + return sg_host_get_speed(host); +} + /** @brief Return the speed of the processor (in flop/s) at a given pstate. See also @ref plugin_energy. * * @param host host to test @@ -579,7 +569,7 @@ void sg_host_set_property_value(sg_host_t host, const char* name, const char* va * @param to where to * @param links [OUT] where to store the list of links (must exist, cannot be nullptr). */ -void sg_host_route(const_sg_host_t from, const_sg_host_t to, xbt_dynar_t links) +void sg_host_get_route(const_sg_host_t from, const_sg_host_t to, xbt_dynar_t links) { std::vector vlinks; from->route_to(to, vlinks, nullptr); @@ -592,7 +582,7 @@ void sg_host_route(const_sg_host_t from, const_sg_host_t to, xbt_dynar_t links) * @param from where from * @param to where to */ -double sg_host_route_latency(const_sg_host_t from, const_sg_host_t to) +double sg_host_get_route_latency(const_sg_host_t from, const_sg_host_t to) { std::vector vlinks; double res = 0; @@ -605,7 +595,7 @@ double sg_host_route_latency(const_sg_host_t from, const_sg_host_t to) * @param from where from * @param to where to */ -double sg_host_route_bandwidth(const_sg_host_t from, const_sg_host_t to) +double sg_host_get_route_bandwidth(const_sg_host_t from, const_sg_host_t to) { double min_bandwidth = -1.0; @@ -619,9 +609,24 @@ double sg_host_route_bandwidth(const_sg_host_t from, const_sg_host_t to) return min_bandwidth; } -void sg_host_send_to(sg_host_t from, sg_host_t to, double byte_amount) +void sg_host_route(const_sg_host_t from, const_sg_host_t to, xbt_dynar_t links) // XBT_ATTRIB_DEPRECATED_v330 +{ + sg_host_get_route(from, to, links); +} + +double sg_host_route_latency(const_sg_host_t from, const_sg_host_t to) // XBT_ATTRIB_DEPRECATED_v330 +{ + return sg_host_get_route_latency(from, to); +} + +double sg_host_route_bandwidth(const_sg_host_t from, const_sg_host_t to) // XBT_ATTRIB_DEPRECATED_v330 +{ + return sg_host_get_route_bandwidth(from, to); +} + +void sg_host_sendto(sg_host_t from, sg_host_t to, double byte_amount) { - from->send_to(to, byte_amount); + from->sendto(to, byte_amount); } /** @brief Displays debugging information about a host */ @@ -669,7 +674,12 @@ const char* sg_host_self_get_name() return res; } -double sg_host_load(const_sg_host_t host) +double sg_host_get_load(const_sg_host_t host) { return host->get_load(); } + +double sg_host_load(const_sg_host_t host) // XBT_ATTRIB_DEPRECATED_v330 +{ + return sg_host_get_load(host); +}