1 /* Copyright (c) 2009-2011. The SimGrid Team.
2 * All rights reserved. */
4 /* This program is free software; you can redistribute it and/or modify it
5 * under the terms of the license (GNU LGPL) which comes with this package. */
7 #include "surf_private.h"
8 #include "surf/surf_resource.h"
9 #include "maxmin_private.h"
10 #include "simgrid/sg_config.h"
11 #include "surf/cpu_cas01_private.h"
13 /* the model objects for physical machines and virtual machines */
14 surf_model_t surf_cpu_model_pm = NULL;
15 surf_model_t surf_cpu_model_vm = NULL;
17 #undef GENERIC_LMM_ACTION
20 #define GENERIC_LMM_ACTION(action) action->generic_lmm_action
21 #define GENERIC_ACTION(action) GENERIC_LMM_ACTION(action).generic_action
22 #define ACTION_GET_CPU(action) ((surf_action_cpu_Cas01_t) action)->cpu
24 typedef struct surf_action_cpu_cas01 {
25 s_surf_action_lmm_t generic_lmm_action;
26 } s_surf_action_cpu_Cas01_t, *surf_action_cpu_Cas01_t;
28 XBT_LOG_NEW_DEFAULT_SUBCATEGORY(surf_cpu, surf,
29 "Logging specific to the SURF CPU IMPROVED module");
32 cpu_running_action_set_that_does_not_need_being_checked = NULL;
35 void *cpu_cas01_create_resource(const char *name, double power_peak,
37 tmgr_trace_t power_trace,
39 e_surf_resource_state_t state_initial,
40 tmgr_trace_t state_trace,
41 xbt_dict_t cpu_properties,
42 surf_model_t cpu_model)
44 cpu_Cas01_t cpu = NULL;
46 xbt_assert(!surf_cpu_resource_priv(surf_cpu_resource_by_name(name)),
47 "Host '%s' declared several times in the platform file",
49 cpu = (cpu_Cas01_t) surf_resource_new(sizeof(s_cpu_Cas01_t),
52 cpu->power_peak = power_peak;
53 xbt_assert(cpu->power_peak > 0, "Power has to be >0");
54 cpu->power_scale = power_scale;
56 xbt_assert(core > 0, "Invalid number of cores %d", core);
60 tmgr_history_add_trace(history, power_trace, 0.0, 0, cpu);
62 cpu->state_current = state_initial;
65 tmgr_history_add_trace(history, state_trace, 0.0, 0, cpu);
68 lmm_constraint_new(cpu_model->model_private->maxmin_system, cpu,
69 cpu->core * cpu->power_scale * cpu->power_peak);
71 /* Note (hypervisor): we create a constraint object for each CPU core, which
72 * is used for making a contraint problem of CPU affinity.
75 /* At now, we assume that a VM does not have a multicore CPU. */
77 xbt_assert(cpu_model == surf_cpu_model_pm);
79 cpu->constraint_core = xbt_new(lmm_constraint_t, core);
82 for (i = 0; i < core; i++) {
83 /* just for a unique id, never used as a string. */
84 void *cnst_id = bprintf("%s:%lu", name, i);
85 cpu->constraint_core[i] =
86 lmm_constraint_new(cpu_model->model_private->maxmin_system, cnst_id,
87 cpu->power_scale * cpu->power_peak);
91 xbt_lib_set(host_lib, name, SURF_CPU_LEVEL, cpu);
93 return xbt_lib_get_elm_or_null(host_lib, name);;
97 static void parse_cpu_init(sg_platf_host_cbarg_t host)
99 /* This function is called when a platform file is parsed. Physical machines
100 * are defined there. Thus, we use the cpu model object for the physical
102 cpu_cas01_create_resource(host->id,
108 host->state_trace, host->properties,
112 static void cpu_add_traces_cpu(void)
114 xbt_dict_cursor_t cursor = NULL;
115 char *trace_name, *elm;
116 static int called = 0;
121 /* connect all traces relative to hosts */
122 xbt_dict_foreach(trace_connect_list_host_avail, cursor, trace_name, elm) {
123 tmgr_trace_t trace = xbt_dict_get_or_null(traces_set_list, trace_name);
124 cpu_Cas01_t host = surf_cpu_resource_by_name(elm);
126 xbt_assert(host, "Host %s undefined", elm);
127 xbt_assert(trace, "Trace %s undefined", trace_name);
130 tmgr_history_add_trace(history, trace, 0.0, 0, host);
133 xbt_dict_foreach(trace_connect_list_power, cursor, trace_name, elm) {
134 tmgr_trace_t trace = xbt_dict_get_or_null(traces_set_list, trace_name);
135 cpu_Cas01_t host = surf_cpu_resource_by_name(elm);
137 xbt_assert(host, "Host %s undefined", elm);
138 xbt_assert(trace, "Trace %s undefined", trace_name);
141 tmgr_history_add_trace(history, trace, 0.0, 0, host);
145 static void cpu_define_callbacks_cas01()
147 sg_platf_host_add_cb(parse_cpu_init);
148 sg_platf_postparse_add_cb(cpu_add_traces_cpu);
151 static int cpu_resource_used(void *resource)
153 surf_model_t cpu_model = ((surf_resource_t) resource)->model;
155 /* Note (hypervisor): we do not need to look up constraint_core[i] here. Even
156 * when a task is pinned or not, its variable object is always linked to the
157 * basic contraint object.
160 return lmm_constraint_used(cpu_model->model_private->maxmin_system,
161 ((cpu_Cas01_t) resource)->constraint);
164 static double cpu_share_resources_lazy(surf_model_t cpu_model, double now)
166 return generic_share_resources_lazy(now, cpu_model);
169 static double cpu_share_resources_full(surf_model_t cpu_model, double now)
171 s_surf_action_cpu_Cas01_t action;
172 return generic_maxmin_share_resources(cpu_model->states.
174 xbt_swag_offset(action,
177 cpu_model->model_private->maxmin_system, lmm_solve);
180 static void cpu_update_actions_state_lazy(surf_model_t cpu_model, double now, double delta)
182 generic_update_actions_state_lazy(now, delta, cpu_model);
185 static void cpu_update_actions_state_full(surf_model_t cpu_model, double now, double delta)
187 generic_update_actions_state_full(now, delta, cpu_model);
190 static void cpu_update_resource_state(void *id,
191 tmgr_trace_event_t event_type,
192 double value, double date)
194 cpu_Cas01_t cpu = id;
195 lmm_variable_t var = NULL;
196 lmm_element_t elem = NULL;
197 surf_model_t cpu_model = ((surf_resource_t) cpu)->model;
199 surf_watched_hosts();
201 if (event_type == cpu->power_event) {
202 /* TODO (Hypervisor): do the same thing for constraint_core[i] */
203 XBT_CRITICAL("FIXME: add power scaling code also for constraint_core[i]");
206 cpu->power_scale = value;
207 lmm_update_constraint_bound(cpu_model->model_private->maxmin_system, cpu->constraint,
208 cpu->core * cpu->power_scale *
211 TRACE_surf_host_set_power(date, cpu->generic_resource.name,
212 cpu->core * cpu->power_scale *
215 while ((var = lmm_get_var_from_cnst
216 (cpu_model->model_private->maxmin_system, cpu->constraint, &elem))) {
217 surf_action_cpu_Cas01_t action = lmm_variable_id(var);
218 lmm_update_variable_bound(cpu_model->model_private->maxmin_system,
219 GENERIC_LMM_ACTION(action).variable,
220 cpu->power_scale * cpu->power_peak);
222 if (tmgr_trace_event_free(event_type))
223 cpu->power_event = NULL;
224 } else if (event_type == cpu->state_event) {
225 /* TODO (Hypervisor): do the same thing for constraint_core[i] */
226 XBT_CRITICAL("FIXME: add state change code also for constraint_core[i]");
230 cpu->state_current = SURF_RESOURCE_ON;
232 lmm_constraint_t cnst = cpu->constraint;
234 cpu->state_current = SURF_RESOURCE_OFF;
236 while ((var = lmm_get_var_from_cnst(cpu_model->model_private->maxmin_system, cnst, &elem))) {
237 surf_action_t action = lmm_variable_id(var);
239 if (surf_action_state_get(action) == SURF_ACTION_RUNNING ||
240 surf_action_state_get(action) == SURF_ACTION_READY ||
241 surf_action_state_get(action) ==
242 SURF_ACTION_NOT_IN_THE_SYSTEM) {
243 action->finish = date;
244 surf_action_state_set(action, SURF_ACTION_FAILED);
248 if (tmgr_trace_event_free(event_type))
249 cpu->state_event = NULL;
251 XBT_CRITICAL("Unknown event ! \n");
261 * This function formulates a constraint problem that pins a given task to
262 * particular cores. Currently, it is possible to pin a task to an exactly one
263 * specific core. The system links the variable object of the task to the
264 * per-core constraint object.
266 * But, the taskset command on Linux takes a mask value specifying a CPU
267 * affinity setting of a given task. If the mask value is 0x03, the given task
268 * will be executed on the first core (CPU0) or the second core (CPU1) on the
269 * given PM. The schedular will determine appropriate placements of tasks,
270 * considering given CPU affinities and task activities.
272 * How should the system formulate constraint problems for an affinity to
276 static void cpu_action_set_affinity(surf_action_t action, void *cpu, unsigned long mask)
278 lmm_variable_t var_obj = ((surf_action_lmm_t) action)->variable;
280 surf_model_t cpu_model = action->model_obj;
281 xbt_assert(cpu_model->type == SURF_MODEL_TYPE_CPU);
282 cpu_Cas01_t CPU = surf_cpu_resource_priv(cpu);
284 XBT_IN("(%p,%lx)", action, mask);
287 unsigned long nbits = 0;
289 /* FIXME: There is much faster algorithms doing this. */
291 for (i = 0; i < CPU->core; i++) {
292 unsigned long has_affinity = (1UL << i) & mask;
298 XBT_CRITICAL("Do not specify multiple cores for an affinity mask.");
299 XBT_CRITICAL("See the comment in cpu_action_set_affinity().");
307 for (i = 0; i < CPU->core; i++) {
308 XBT_INFO("clear affinity %p to cpu-%lu@%s", action, i, CPU->generic_resource.name);
309 lmm_shrink(cpu_model->model_private->maxmin_system, CPU->constraint_core[i], var_obj);
311 unsigned long has_affinity = (1UL << i) & mask;
313 XBT_INFO("set affinity %p to cpu-%lu@%s", action, i, CPU->generic_resource.name);
314 lmm_expand(cpu_model->model_private->maxmin_system, CPU->constraint_core[i], var_obj, 1.0);
318 if (cpu_model->model_private->update_mechanism == UM_LAZY) {
319 XBT_WARN("FIXME (hypervisor): Do we need to do something for the LAZY mode?");
325 static surf_action_t cpu_execute(void *cpu, double size)
327 surf_action_cpu_Cas01_t action = NULL;
328 cpu_Cas01_t CPU = surf_cpu_resource_priv(cpu);
329 surf_model_t cpu_model = ((surf_resource_t) CPU)->model;
331 XBT_IN("(%s,%g)", surf_resource_name(CPU), size);
333 surf_action_new(sizeof(s_surf_action_cpu_Cas01_t), size,
335 CPU->state_current != SURF_RESOURCE_ON);
337 GENERIC_LMM_ACTION(action).suspended = 0; /* Should be useless because of the
338 calloc but it seems to help valgrind... */
340 /* Note (hypervisor): here, the bound value of the variable is set to the
341 * capacity of a CPU core. But, after MSG_{task/vm}_set_bound() were added to
342 * the hypervisor branch, this bound value is overwritten in
343 * SIMIX_host_execute().
344 * TODO: cleanup this.
346 GENERIC_LMM_ACTION(action).variable =
347 lmm_variable_new(cpu_model->model_private->maxmin_system, action,
348 GENERIC_ACTION(action).priority,
349 CPU->power_scale * CPU->power_peak, 1 + CPU->core); // the basic constraint plus core-specific constraints
350 if (cpu_model->model_private->update_mechanism == UM_LAZY) {
351 GENERIC_LMM_ACTION(action).index_heap = -1;
352 GENERIC_LMM_ACTION(action).last_update = surf_get_clock();
353 GENERIC_LMM_ACTION(action).last_value = 0.0;
355 lmm_expand(cpu_model->model_private->maxmin_system, CPU->constraint,
356 GENERIC_LMM_ACTION(action).variable, 1.0);
358 return (surf_action_t) action;
361 static surf_action_t cpu_action_sleep(void *cpu, double duration)
363 surf_action_cpu_Cas01_t action = NULL;
364 cpu_Cas01_t CPU = surf_cpu_resource_priv(cpu);
365 surf_model_t cpu_model = ((surf_resource_t) CPU)->model;
368 duration = MAX(duration, MAXMIN_PRECISION);
370 XBT_IN("(%s,%g)", surf_resource_name(surf_cpu_resource_priv(cpu)), duration);
371 action = (surf_action_cpu_Cas01_t) cpu_execute(cpu, 1.0);
372 // FIXME: sleep variables should not consume 1.0 in lmm_expand
373 GENERIC_ACTION(action).max_duration = duration;
374 GENERIC_LMM_ACTION(action).suspended = 2;
375 if (duration == NO_MAX_DURATION) {
376 /* Move to the *end* of the corresponding action set. This convention
377 is used to speed up update_resource_state */
378 xbt_swag_remove(action, ((surf_action_t) action)->state_set);
379 ((surf_action_t) action)->state_set =
380 cpu_running_action_set_that_does_not_need_being_checked;
381 xbt_swag_insert(action, ((surf_action_t) action)->state_set);
384 lmm_update_variable_weight(cpu_model->model_private->maxmin_system,
385 GENERIC_LMM_ACTION(action).variable, 0.0);
386 if (cpu_model->model_private->update_mechanism == UM_LAZY) { // remove action from the heap
387 surf_action_lmm_heap_remove(cpu_model->model_private->action_heap,(surf_action_lmm_t)action);
388 // this is necessary for a variable with weight 0 since such
389 // variables are ignored in lmm and we need to set its max_duration
390 // correctly at the next call to share_resources
391 xbt_swag_insert_at_head(action, cpu_model->model_private->modified_set);
395 return (surf_action_t) action;
398 static e_surf_resource_state_t cpu_get_state(void *cpu)
400 return ((cpu_Cas01_t)surf_cpu_resource_priv(cpu))->state_current;
403 static void cpu_set_state(void *cpu, e_surf_resource_state_t state)
405 ((cpu_Cas01_t)surf_cpu_resource_priv(cpu))->state_current = state;
408 static double cpu_get_speed(void *cpu, double load)
410 return load * ((cpu_Cas01_t)surf_cpu_resource_priv(cpu))->power_peak;
413 static int cpu_get_core(void *cpu)
415 return ((cpu_Cas01_t)surf_cpu_resource_priv(cpu))->core;
417 static double cpu_get_available_speed(void *cpu)
419 /* number between 0 and 1 */
420 return ((cpu_Cas01_t)surf_cpu_resource_priv(cpu))->power_scale;
423 static void cpu_finalize(surf_model_t cpu_model)
425 lmm_system_free(cpu_model->model_private->maxmin_system);
426 cpu_model->model_private->maxmin_system = NULL;
428 if (cpu_model->model_private->action_heap)
429 xbt_heap_free(cpu_model->model_private->action_heap);
430 xbt_swag_free(cpu_model->model_private->modified_set);
432 surf_model_exit(cpu_model);
435 xbt_swag_free(cpu_running_action_set_that_does_not_need_being_checked);
436 cpu_running_action_set_that_does_not_need_being_checked = NULL;
439 static surf_model_t surf_cpu_model_init_cas01(void)
441 s_surf_action_t action;
442 s_surf_action_cpu_Cas01_t comp;
444 char *optim = xbt_cfg_get_string(_sg_cfg_set, "cpu/optim");
446 xbt_cfg_get_boolean(_sg_cfg_set, "cpu/maxmin_selective_update");
448 surf_model_t cpu_model = surf_model_init();
450 if (!strcmp(optim, "Full")) {
451 cpu_model->model_private->update_mechanism = UM_FULL;
452 cpu_model->model_private->selective_update = select;
453 } else if (!strcmp(optim, "Lazy")) {
454 cpu_model->model_private->update_mechanism = UM_LAZY;
455 cpu_model->model_private->selective_update = 1;
456 xbt_assert((select == 1)
458 (xbt_cfg_is_default_value
459 (_sg_cfg_set, "cpu/maxmin_selective_update")),
460 "Disabling selective update while using the lazy update mechanism is dumb!");
462 xbt_die("Unsupported optimization (%s) for this model", optim);
465 cpu_running_action_set_that_does_not_need_being_checked =
466 xbt_swag_new(xbt_swag_offset(action, state_hookup));
468 cpu_model->name = "cpu";
469 cpu_model->type = SURF_MODEL_TYPE_CPU;
471 cpu_model->action_unref = surf_action_unref;
472 cpu_model->action_cancel = surf_action_cancel;
473 cpu_model->action_state_set = surf_action_state_set;
475 cpu_model->model_private->resource_used = cpu_resource_used;
477 if (cpu_model->model_private->update_mechanism == UM_LAZY) {
478 cpu_model->model_private->share_resources =
479 cpu_share_resources_lazy;
480 cpu_model->model_private->update_actions_state =
481 cpu_update_actions_state_lazy;
482 } else if (cpu_model->model_private->update_mechanism == UM_FULL) {
483 cpu_model->model_private->share_resources =
484 cpu_share_resources_full;
485 cpu_model->model_private->update_actions_state =
486 cpu_update_actions_state_full;
488 xbt_die("Invalid cpu update mechanism!");
490 cpu_model->model_private->update_resource_state =
491 cpu_update_resource_state;
492 cpu_model->model_private->finalize = cpu_finalize;
494 cpu_model->suspend = surf_action_suspend;
495 cpu_model->resume = surf_action_resume;
496 cpu_model->is_suspended = surf_action_is_suspended;
497 cpu_model->set_max_duration = surf_action_set_max_duration;
498 cpu_model->set_priority = surf_action_set_priority;
499 cpu_model->set_bound = surf_action_set_bound;
500 cpu_model->set_affinity = cpu_action_set_affinity;
502 cpu_model->set_category = surf_action_set_category;
504 cpu_model->get_remains = surf_action_get_remains;
506 cpu_model->extension.cpu.execute = cpu_execute;
507 cpu_model->extension.cpu.sleep = cpu_action_sleep;
509 cpu_model->extension.cpu.get_state = cpu_get_state;
510 cpu_model->extension.cpu.set_state = cpu_set_state;
511 cpu_model->extension.cpu.get_core = cpu_get_core;
512 cpu_model->extension.cpu.get_speed = cpu_get_speed;
513 cpu_model->extension.cpu.get_available_speed =
514 cpu_get_available_speed;
515 cpu_model->extension.cpu.add_traces = cpu_add_traces_cpu;
517 if (!cpu_model->model_private->maxmin_system) {
518 cpu_model->model_private->maxmin_system = lmm_system_new(cpu_model->model_private->selective_update);
520 if (cpu_model->model_private->update_mechanism == UM_LAZY) {
521 cpu_model->model_private->action_heap = xbt_heap_new(8, NULL);
522 xbt_heap_set_update_callback(cpu_model->model_private->action_heap,
523 surf_action_lmm_update_index_heap);
524 cpu_model->model_private->modified_set =
525 xbt_swag_new(xbt_swag_offset(comp, generic_lmm_action.action_list_hookup));
526 cpu_model->model_private->maxmin_system->keep_track = cpu_model->model_private->modified_set;
532 /*********************************************************************/
533 /* Basic sharing model for CPU: that is where all this started... ;) */
534 /*********************************************************************/
535 /* @InProceedings{casanova01simgrid, */
536 /* author = "H. Casanova", */
537 /* booktitle = "Proceedings of the IEEE Symposium on Cluster Computing */
538 /* and the Grid (CCGrid'01)", */
539 /* publisher = "IEEE Computer Society", */
540 /* title = "Simgrid: {A} Toolkit for the Simulation of Application */
544 /* note = "Available at */
545 /* \url{http://grail.sdsc.edu/papers/simgrid_ccgrid01.ps.gz}." */
549 void surf_cpu_model_init_Cas01(void)
551 char *optim = xbt_cfg_get_string(_sg_cfg_set, "cpu/optim");
553 xbt_assert(!surf_cpu_model_pm);
554 xbt_assert(!surf_cpu_model_vm);
556 if (strcmp(optim, "TI") == 0) {
557 /* FIXME: do we have to supprot TI? for VM */
558 surf_cpu_model_pm = surf_cpu_model_init_ti();
559 XBT_INFO("TI model is used (it will crashed since this is the hypervisor branch)");
561 surf_cpu_model_pm = surf_cpu_model_init_cas01();
562 surf_cpu_model_vm = surf_cpu_model_init_cas01();
564 /* cpu_model is registered only to model_list, and not to
565 * model_list_invoke. The shared_resource callback function will be called
566 * from that of the workstation model. */
567 xbt_dynar_push(model_list, &surf_cpu_model_pm);
568 xbt_dynar_push(model_list, &surf_cpu_model_vm);
570 cpu_define_callbacks_cas01();
574 /* TODO: do we address nested virtualization later? */
576 surf_model_t cpu_model_cas01(int level){
577 // TODO this table should be allocated
578 if(!surf_cpu_model[level])
580 return surf_cpu_model[level];