model-net-lp.c 22.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11
/*
 * Copyright (C) 2014 University of Chicago.
 * See COPYRIGHT notice in top-level directory.
 *
 */

#include <stddef.h>
#include <assert.h>
#include "codes/model-net.h"
#include "codes/model-net-method.h"
#include "codes/model-net-lp.h"
12
#include "codes/model-net-sched.h"
13 14 15 16 17 18 19 20 21 22 23 24 25
#include "codes/codes_mapping.h"
#include "codes/jenkins-hash.h"

#define MN_NAME "model_net_base"

/**** BEGIN SIMULATION DATA STRUCTURES ****/

int model_net_base_magic;

// message-type specific offsets - don't want to get bitten later by alignment
// issues...
static int msg_offsets[MAX_NETS];

26 27
typedef struct model_net_base_params_s {
    model_net_sched_cfg_params sched_params;
28
    uint64_t packet_size;
29
    int use_recv_queue;
30 31
} model_net_base_params;

32
/* annotation-specific parameters (unannotated entry occurs at the
33 34 35 36 37
 * last index) */
static int                       num_params = 0;
static const char              * annos[CONFIGURATION_MAX_ANNOS];
static model_net_base_params     all_params[CONFIGURATION_MAX_ANNOS];

38 39
static tw_stime mn_sample_interval = 0.0;
static int mn_sample_enabled = 0;
40
static tw_stime mn_sample_end = 0.0;
41

42 43
typedef struct model_net_base_state {
    int net_id;
44
    // whether scheduler loop is running
45
    int in_sched_send_loop, in_sched_recv_loop;
46 47 48
    // unique message id counter. This doesn't get decremented on RC to prevent
    // optimistic orderings using "stale" ids
    uint64_t msg_id;
49 50
    // model-net schedulers
    model_net_sched *sched_send, *sched_recv;
51 52
    // parameters
    const model_net_base_params * params;
53 54 55 56 57 58
    // lp type and state of underlying model net method - cache here so we
    // don't have to constantly look up
    const tw_lptype *sub_type;
    void *sub_state;
} model_net_base_state;

59

60 61 62 63
/**** END SIMULATION DATA STRUCTURES ****/

/**** BEGIN LP, EVENT PROCESSING FUNCTION DECLS ****/

64
/* ROSS LP processing functions */
65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105
static void model_net_base_lp_init(
        model_net_base_state * ns,
        tw_lp * lp);
static void model_net_base_event(
        model_net_base_state * ns,
        tw_bf * b,
        model_net_wrap_msg * m,
        tw_lp * lp);
static void model_net_base_event_rc(
        model_net_base_state * ns,
        tw_bf * b,
        model_net_wrap_msg * m,
        tw_lp * lp);
static void model_net_base_finalize(
        model_net_base_state * ns,
        tw_lp * lp);

/* event type handlers */
static void handle_new_msg(
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp);
static void handle_sched_next(
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp);
static void handle_new_msg_rc(
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp);
static void handle_sched_next_rc(
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp);

/* ROSS function pointer table for this LP */
tw_lptype model_net_base_lp = {
106 107 108 109
    (init_f) model_net_base_lp_init,
    (pre_run_f) NULL,
    (event_f) model_net_base_event,
    (revent_f) model_net_base_event_rc,
110 111
    (commit_f) NULL,
    (final_f)  model_net_base_finalize,
112 113
    (map_f) codes_mapping,
    sizeof(model_net_base_state),
114 115
};

116 117 118 119 120 121 122 123 124 125 126 127 128
void mn_event_collect(model_net_wrap_msg *m, tw_lp *lp, char *buffer)
{
    int type = (int) m->h.event_type;
    memcpy(buffer, &type, sizeof(type));
}

st_event_collect mn_event_types = {
    (rbev_col_f) mn_event_collect,
     sizeof(int),
     (ev_col_f) mn_event_collect,
     sizeof(int),
};

129 130 131 132
/**** END LP, EVENT PROCESSING FUNCTION DECLS ****/

/**** BEGIN IMPLEMENTATIONS ****/

133
void model_net_enable_sampling(tw_stime interval, tw_stime end)
134 135
{
    mn_sample_interval = interval;
136
    mn_sample_end = end;
137 138 139 140 141 142 143 144 145 146 147
    mn_sample_enabled = 1;
}

int model_net_sampling_enabled(void)
{
    return mn_sample_enabled;
}

// schedule sample event - want to be precise, so no noise here
static void issue_sample_event(tw_lp *lp)
{
148 149 150 151 152 153
    if (tw_now(lp) + mn_sample_interval < mn_sample_end + 0.0001) {
        tw_event *e = tw_event_new(lp->gid, mn_sample_interval, lp);
        model_net_wrap_msg *m = tw_event_data(e);
        msg_set_header(model_net_base_magic, MN_BASE_SAMPLE, lp->gid, &m->h);
        tw_event_send(e);
    }
154 155
}

156 157 158 159
void model_net_base_register(int *do_config_nets){
    // here, we initialize ALL lp types to use the base type
    for (int i = 0; i < MAX_NETS; i++){
        if (do_config_nets[i]){
160 161 162 163 164 165 166
            // some model-net lps need custom registration hooks (dragonfly).
            // Those that don't NULL out the reg. function
            if (method_array[i]->mn_register == NULL)
                lp_type_register(model_net_lp_config_names[i],
                        &model_net_base_lp);
            else
                method_array[i]->mn_register(&model_net_base_lp);
167 168 169 170 171 172 173
            if (g_st_ev_rb_collect || g_st_ev_collect)
            {
                if (method_array[i]->mn_ev_register == NULL)
                    ev_type_register(model_net_lp_config_names[i], &mn_event_types);
                else
                    method_array[i]->mn_ev_register(&mn_event_types);
            }
174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193
        }
    }
}

static void base_read_config(const char * anno, model_net_base_params *p){
    char sched[MAX_NAME_LENGTH];
    long int packet_size_l = 0;
    uint64_t packet_size;
    int ret;

    ret = configuration_get_value(&config, "PARAMS", "modelnet_scheduler",
            anno, sched, MAX_NAME_LENGTH);
    configuration_get_value_longint(&config, "PARAMS", "packet_size", anno,
            &packet_size_l);
    packet_size = packet_size_l;

    if (ret > 0){
        int i;
        for (i = 0; i < MAX_SCHEDS; i++){
            if (strcmp(sched_names[i], sched) == 0){
194
                p->sched_params.type = i;
195 196 197 198 199
                break;
            }
        }
        if (i == MAX_SCHEDS){
            tw_error(TW_LOC,"Unknown value for PARAMS:modelnet-scheduler : "
200
                    "%s", sched);
201 202 203 204
        }
    }
    else{
        // default: FCFS
205
        p->sched_params.type = MN_SCHED_FCFS;
206 207
    }

208 209
    // get scheduler-specific parameters
    if (p->sched_params.type == MN_SCHED_PRIO){
210
        // prio scheduler uses default parameters
211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240
        int             * num_prios = &p->sched_params.u.prio.num_prios;
        enum sched_type * sub_stype = &p->sched_params.u.prio.sub_stype;
        // number of priorities to allocate
        ret = configuration_get_value_int(&config, "PARAMS",
                "prio-sched-num-prios", anno, num_prios);
        if (ret != 0)
            *num_prios = 10;

        ret = configuration_get_value(&config, "PARAMS",
                "prio-sched-sub-sched", anno, sched, MAX_NAME_LENGTH);
        if (ret == 0)
            *sub_stype = MN_SCHED_FCFS;
        else{
            int i;
            for (i = 0; i < MAX_SCHEDS; i++){
                if (strcmp(sched_names[i], sched) == 0){
                    *sub_stype = i;
                    break;
                }
            }
            if (i == MAX_SCHEDS){
                tw_error(TW_LOC, "Unknown value for "
                        "PARAMS:prio-sched-sub-sched %s", sched);
            }
            else if (i == MN_SCHED_PRIO){
                tw_error(TW_LOC, "priority scheduler cannot be used as a "
                        "priority scheduler's sub sched "
                        "(PARAMS:prio-sched-sub-sched)");
            }
        }
241 242
    }

243 244 245 246 247 248 249 250 251 252 253 254 255
    if (p->sched_params.type == MN_SCHED_FCFS_FULL ||
            (p->sched_params.type == MN_SCHED_PRIO &&
             p->sched_params.u.prio.sub_stype == MN_SCHED_FCFS_FULL)){
        // override packet size to something huge (leave a bit in the unlikely
        // case that an op using packet size causes overflow)
        packet_size = 1ull << 62;
    }
    else if (!packet_size &&
            (p->sched_params.type != MN_SCHED_FCFS_FULL ||
             (p->sched_params.type == MN_SCHED_PRIO &&
              p->sched_params.u.prio.sub_stype != MN_SCHED_FCFS_FULL))){
        packet_size = 512;
        fprintf(stderr, "WARNING, no packet size specified, setting packet "
Jonathan Jenkins's avatar
Jonathan Jenkins committed
256
                "size to %llu\n", LLU(packet_size));
257 258 259
    }


260 261 262 263
    p->packet_size = packet_size;
}

void model_net_base_configure(){
264 265 266 267 268
    uint32_t h1=0, h2=0;

    bj_hashlittle2(MN_NAME, strlen(MN_NAME), &h1, &h2);
    model_net_base_magic = h1+h2;

269 270 271
    // set up offsets - doesn't matter if they are actually used or not
    msg_offsets[SIMPLENET] =
        offsetof(model_net_wrap_msg, msg.m_snet);
Jonathan Jenkins's avatar
Jonathan Jenkins committed
272 273
    msg_offsets[SIMPLEP2P] =
        offsetof(model_net_wrap_msg, msg.m_sp2p);
274 275 276 277
    msg_offsets[TORUS] =
        offsetof(model_net_wrap_msg, msg.m_torus);
    msg_offsets[DRAGONFLY] =
        offsetof(model_net_wrap_msg, msg.m_dfly);
278 279 280
    // note: dragonfly router uses the same event struct
    msg_offsets[DRAGONFLY_ROUTER] =
        offsetof(model_net_wrap_msg, msg.m_dfly);
281 282 283 284
    msg_offsets[DRAGONFLY_CUSTOM] =
        offsetof(model_net_wrap_msg, msg.m_dfly);
    msg_offsets[DRAGONFLY_CUSTOM_ROUTER] =
        offsetof(model_net_wrap_msg, msg.m_dfly);
285 286
    msg_offsets[SLIMFLY] =
        offsetof(model_net_wrap_msg, msg.m_slim);
287 288
    msg_offsets[FATTREE] =
	offsetof(model_net_wrap_msg, msg.m_fat);
289 290
    msg_offsets[LOGGP] =
        offsetof(model_net_wrap_msg, msg.m_loggp);
291

292 293 294 295 296 297 298 299 300 301
    // perform the configuration(s)
    // This part is tricky, as we basically have to look up all annotations that
    // have LP names of the form modelnet_*. For each of those, we need to read
    // the base parameters
    // - the init is a little easier as we can use the LP-id to look up the
    // annotation

    // first grab all of the annotations and store locally
    for (int c = 0; c < lpconf.lpannos_count; c++){
        const config_anno_map_t *amap = &lpconf.lpannos[c];
302
        if (strncmp("modelnet_", amap->lp_name.ptr, 9) == 0){
303 304 305
            for (int n = 0; n < amap->num_annos; n++){
                int a;
                for (a = 0; a < num_params; a++){
306 307
                    if (annos[a] != NULL && amap->annotations[n].ptr != NULL &&
                            strcmp(amap->annotations[n].ptr, annos[a]) == 0){
308 309 310 311 312
                        break;
                    }
                }
                if (a == num_params){
                    // found a new annotation
313
                    annos[num_params++] = amap->annotations[n].ptr;
314 315 316 317 318 319 320 321 322 323 324 325 326
                }
            }
            if (amap->has_unanno_lp){
                int a;
                for (a = 0; a < num_params; a++){
                    if (annos[a] == NULL)
                        break;
                }
                if (a == num_params){
                    // found a new (empty) annotation
                    annos[num_params++] = NULL;
                }
            }
327 328
        }
    }
329 330 331 332 333 334

    // now that we have all of the annos for all of the networks, loop through
    // and read the configs
    for (int i = 0; i < num_params; i++){
        base_read_config(annos[i], &all_params[i]);
    }
335 336 337 338 339 340
}

void model_net_base_lp_init(
        model_net_base_state * ns,
        tw_lp * lp){
    // obtain the underlying lp type through codes-mapping
341
    char lp_type_name[MAX_NAME_LENGTH], anno[MAX_NAME_LENGTH];
342 343
    int dummy;

344
    codes_mapping_get_lp_info(lp->gid, NULL, &dummy,
345 346
            lp_type_name, &dummy, anno, &dummy, &dummy);

347 348
    ns->msg_id = 0;

349 350 351 352 353 354 355 356
    // get annotation-specific parameters
    for (int i = 0; i < num_params; i++){
        if ((anno[0]=='\0' && annos[i] == NULL) ||
                strcmp(anno, annos[i]) == 0){
            ns->params = &all_params[i];
            break;
        }
    }
357 358 359 360 361 362 363 364 365

    // find the corresponding method name / index
    for (int i = 0; i < MAX_NETS; i++){
        if (strcmp(model_net_lp_config_names[i], lp_type_name) == 0){
            ns->net_id = i;
            break;
        }
    }

366 367
    ns->sched_send = malloc(sizeof(model_net_sched));
    ns->sched_recv = malloc(sizeof(model_net_sched));
368
    // init both the sender queue and the 'receiver' queue
369 370 371 372
    model_net_sched_init(&ns->params->sched_params, 0, method_array[ns->net_id],
            ns->sched_send);
    model_net_sched_init(&ns->params->sched_params, 1, method_array[ns->net_id],
            ns->sched_recv);
373

374 375 376 377 378 379 380
    ns->sub_type = model_net_get_lp_type(ns->net_id);
    // NOTE: some models actually expect LP state to be 0 initialized...
    // *cough anything that uses mn_stats_array cough*
    ns->sub_state = calloc(1, ns->sub_type->state_sz);

    // initialize the model-net method
    ns->sub_type->init(ns->sub_state, lp);
381 382 383 384 385 386

    // check validity of sampling function
    event_f  sample  = method_array[ns->net_id]->mn_sample_fn;
    revent_f rsample = method_array[ns->net_id]->mn_sample_rc_fn;
    if (model_net_sampling_enabled()) {
        if (sample == NULL) {
387 388 389
            /* MM: Commented out temporarily--- */
            //tw_error(TW_LOC,
            //        "Sampling requested for a model that doesn't provide it\n");
390 391 392 393
        }
        else if (rsample == NULL &&
                (g_tw_synchronization_protocol == OPTIMISTIC ||
                 g_tw_synchronization_protocol == OPTIMISTIC_DEBUG)) {
394 395 396
            /* MM: Commented out temporarily--- */
            //tw_error(TW_LOC,
            //        "Sampling requested for a model that doesn't provide it\n");
397 398 399 400 401 402 403 404
        }
        else {
            init_f sinit = method_array[ns->net_id]->mn_sample_init_fn;
            if (sinit != NULL)
                sinit(ns->sub_state, lp);
            issue_sample_event(lp);
        }
    }
405 406 407 408 409 410 411
}

void model_net_base_event(
        model_net_base_state * ns,
        tw_bf * b,
        model_net_wrap_msg * m,
        tw_lp * lp){
412 413 414 415

    if(m->h.magic != model_net_base_magic)
        printf("\n LP ID mismatched %d ", lp->gid);

416
    assert(m->h.magic == model_net_base_magic);
417 418

    void * sub_msg;
419
    switch (m->h.event_type){
420 421 422 423 424 425
        case MN_BASE_NEW_MSG:
            handle_new_msg(ns, b, m, lp);
            break;
        case MN_BASE_SCHED_NEXT:
            handle_sched_next(ns, b, m, lp);
            break;
426 427 428 429 430 431 432
        case MN_BASE_SAMPLE: ;
            event_f sample = method_array[ns->net_id]->mn_sample_fn;
            assert(model_net_sampling_enabled() && sample != NULL);
            sub_msg = ((char*)m)+msg_offsets[ns->net_id];
            sample(ns->sub_state, b, sub_msg, lp);
            issue_sample_event(lp);
            break;
433
        case MN_BASE_PASS: ;
434
            sub_msg = ((char*)m)+msg_offsets[ns->net_id];
435 436 437 438 439 440 441 442 443 444 445 446 447 448
            ns->sub_type->event(ns->sub_state, b, sub_msg, lp);
            break;
        /* ... */
        default:
            assert(!"model_net_base event type not known");
            break;
    }
}

void model_net_base_event_rc(
        model_net_base_state * ns,
        tw_bf * b,
        model_net_wrap_msg * m,
        tw_lp * lp){
449
    assert(m->h.magic == model_net_base_magic);
450 451

    void * sub_msg;
452
    switch (m->h.event_type){
453 454 455 456 457 458
        case MN_BASE_NEW_MSG:
            handle_new_msg_rc(ns, b, m, lp);
            break;
        case MN_BASE_SCHED_NEXT:
            handle_sched_next_rc(ns, b, m, lp);
            break;
459 460 461 462 463 464
        case MN_BASE_SAMPLE: ;
            revent_f sample_rc = method_array[ns->net_id]->mn_sample_rc_fn;
            assert(model_net_sampling_enabled() && sample_rc != NULL);
            sub_msg = ((char*)m)+msg_offsets[ns->net_id];
            sample_rc(ns->sub_state, b, sub_msg, lp);
            break;
465
        case MN_BASE_PASS: ;
466
            sub_msg = ((char*)m)+msg_offsets[ns->net_id];
467 468 469 470 471 472 473 474 475 476 477 478
            ns->sub_type->revent(ns->sub_state, b, sub_msg, lp);
            break;
        /* ... */
        default:
            assert(!"model_net_base event type not known");
            break;
    }
}

void model_net_base_finalize(
        model_net_base_state * ns,
        tw_lp * lp){
479 480 481
    final_f sfini = method_array[ns->net_id]->mn_sample_fini_fn;
    if (sfini != NULL)
        sfini(ns->sub_state, lp);
482 483 484 485
    ns->sub_type->final(ns->sub_state, lp);
    free(ns->sub_state);
}

486
/// bitfields used:
487
/// c31 - we initiated a sched_next event
488 489 490 491 492
void handle_new_msg(
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp){
493
    // simply pass down to the scheduler
494
    model_net_request *r = &m->msg.m_base.req;
495 496
    // don't forget to set packet size, now that we're responsible for it!
    r->packet_size = ns->params->packet_size;
497
    r->msg_id = ns->msg_id++;
498 499
    void * m_data = m+1;
    void *remote = NULL, *local = NULL;
500
    if (r->remote_event_size > 0){
501 502
        remote = m_data;
        m_data = (char*)m_data + r->remote_event_size;
503 504
    }
    if (r->self_event_size > 0){
505
        local = m_data;
506
    }
507

508
    // set message-specific params
509 510
    int is_from_remote = m->msg.m_base.is_from_remote;
    model_net_sched *ss = is_from_remote ? ns->sched_recv : ns->sched_send;
511
    int *in_sched_loop = is_from_remote  ?
512 513 514
        &ns->in_sched_recv_loop : &ns->in_sched_send_loop;
    model_net_sched_add(r, &m->msg.m_base.sched_params, r->remote_event_size,
            remote, r->self_event_size, local, ss, &m->msg.m_base.rc, lp);
515

516
    if (*in_sched_loop == 0){
517 518
        b->c31 = 1;
        /* No need to issue an extra sched-next event if we're currently idle */
519
        *in_sched_loop = 1;
520 521 522 523 524
        /* NOTE: we can do this because the sched rc struct in the event is
         * *very* lightly used (there's harmless overlap in usage for the
         * priority scheduler) */
        handle_sched_next(ns, b, m, lp);
        assert(*in_sched_loop); // we shouldn't have fallen out of the loop
525 526
    }
}
527 528 529

void handle_new_msg_rc(
        model_net_base_state *ns,
530
        tw_bf *b,
531 532
        model_net_wrap_msg *m,
        tw_lp *lp){
533 534
    int is_from_remote = m->msg.m_base.is_from_remote;
    model_net_sched *ss = is_from_remote ? ns->sched_recv : ns->sched_send;
535
    int *in_sched_loop = is_from_remote  ?
536 537
        &ns->in_sched_recv_loop : &ns->in_sched_send_loop;

538 539
    if (b->c31) {
        handle_sched_next_rc(ns, b, m, lp);
540
        *in_sched_loop = 0;
541
    }
542
    model_net_sched_add_rc(ss, &m->msg.m_base.rc, lp);
543
}
544 545 546 547

/// bitfields used
/// c0 - scheduler loop is finished
void handle_sched_next(
548 549 550 551
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp){
552
    tw_stime poffset;
553 554 555 556 557
    int is_from_remote = m->msg.m_base.is_from_remote;
    model_net_sched * ss = is_from_remote ? ns->sched_recv : ns->sched_send;
    int *in_sched_loop = is_from_remote ?
        &ns->in_sched_recv_loop : &ns->in_sched_send_loop;
    int ret = model_net_sched_next(&poffset, ss, m+1, &m->msg.m_base.rc, lp);
558 559 560 561
    // we only need to know whether scheduling is finished or not - if not,
    // go to the 'next iteration' of the loop
    if (ret == -1){
        b->c0 = 1;
562
        *in_sched_loop = 0;
563
    }
564 565 566
    // Currently, only a subset of the network implementations use the
    // callback-based scheduling loop (model_net_method_idle_event).
    // For all others, we need to schedule the next packet
567
    // immediately
Jonathan Jenkins's avatar
Jonathan Jenkins committed
568
    else if (ns->net_id == SIMPLEP2P || ns->net_id == TORUS){
569
        tw_event *e = tw_event_new(lp->gid,
570
                poffset+codes_local_latency(lp), lp);
571
        model_net_wrap_msg *m_wrap = tw_event_data(e);
572
        msg_set_header(model_net_base_magic, MN_BASE_SCHED_NEXT, lp->gid,
573 574
                &m_wrap->h);
        m_wrap->msg.m_base.is_from_remote = is_from_remote;
575 576
        // no need to set m_base here
        tw_event_send(e);
577 578
    }
}
579

580 581 582 583 584
void handle_sched_next_rc(
        model_net_base_state * ns,
        tw_bf *b,
        model_net_wrap_msg * m,
        tw_lp * lp){
585 586 587 588
    int is_from_remote = m->msg.m_base.is_from_remote;
    model_net_sched * ss = is_from_remote ? ns->sched_recv : ns->sched_send;
    int *in_sched_loop = is_from_remote ?
        &ns->in_sched_recv_loop : &ns->in_sched_send_loop;
589

590
    model_net_sched_next_rc(ss, m+1, &m->msg.m_base.rc, lp);
591
    if (b->c0){
592
        *in_sched_loop = 1;
593
    }
Jonathan Jenkins's avatar
Jonathan Jenkins committed
594
    else if (ns->net_id == SIMPLEP2P || ns->net_id == TORUS){
595 596
        codes_local_latency_reverse(lp);
    }
597 598 599 600 601 602 603 604 605 606 607 608 609
}

/**** END IMPLEMENTATIONS ****/

tw_event * model_net_method_event_new(
        tw_lpid dest_gid,
        tw_stime offset_ts,
        tw_lp *sender,
        int net_id,
        void **msg_data,
        void **extra_data){
    tw_event *e = tw_event_new(dest_gid, offset_ts, sender);
    model_net_wrap_msg *m_wrap = tw_event_data(e);
610 611
    msg_set_header(model_net_base_magic, MN_BASE_PASS, sender->gid,
            &m_wrap->h);
612 613 614 615 616 617 618 619
    *msg_data = ((char*)m_wrap)+msg_offsets[net_id];
    // extra_data is optional
    if (extra_data != NULL){
        *extra_data = m_wrap + 1;
    }
    return e;
}

620 621 622 623 624 625 626 627 628 629 630 631
void model_net_method_send_msg_recv_event(
        tw_lpid final_dest_lp,
        tw_lpid dest_mn_lp,
        tw_lpid src_lp, // the "actual" source (as opposed to the model net lp)
        uint64_t msg_size,
        int is_pull,
        uint64_t pull_size,
        int remote_event_size,
        const mn_sched_params *sched_params,
        const char * category,
        int net_id,
        void * msg,
632
        tw_stime offset,
633
        tw_lp *sender){
634
    tw_event *e =
635
        tw_event_new(dest_mn_lp, offset+codes_local_latency(sender), sender);
636 637 638 639 640 641 642 643
    model_net_wrap_msg *m = tw_event_data(e);
    msg_set_header(model_net_base_magic, MN_BASE_NEW_MSG, sender->gid, &m->h);

    if (sched_params != NULL)
        m->msg.m_base.sched_params = *sched_params;
    else
        model_net_sched_set_default_params(&m->msg.m_base.sched_params);

644 645 646
    model_net_request *r = &m->msg.m_base.req;
    r->final_dest_lp = final_dest_lp;
    r->src_lp = src_lp;
647
    // for "recv" events, set the "dest" to this LP in the case of a pull event
648 649 650 651 652 653 654 655 656
    r->dest_mn_lp = sender->gid;
    r->pull_size = pull_size;
    r->msg_size = msg_size;
    // TODO: document why we're setting packet_size this way
    r->packet_size = msg_size;
    r->net_id = net_id;
    r->is_pull = is_pull;
    r->remote_event_size = remote_event_size;
    r->self_event_size = 0;
657 658
    m->msg.m_base.is_from_remote = 1;

659 660
    strncpy(r->category, category, CATEGORY_NAME_MAX-1);
    r->category[CATEGORY_NAME_MAX-1] = '\0';
661 662 663 664 665 666 667 668 669

    if (remote_event_size > 0){
        void * m_dat = model_net_method_get_edata(net_id, msg);
        memcpy(m+1, m_dat, remote_event_size);
    }

    tw_event_send(e);
}

Jonathan Jenkins's avatar
Jonathan Jenkins committed
670 671 672 673
void model_net_method_send_msg_recv_event_rc(tw_lp *sender){
    codes_local_latency_reverse(sender);
}

674 675 676

void model_net_method_idle_event(tw_stime offset_ts, int is_recv_queue,
        tw_lp * lp){
677 678
    tw_event *e = tw_event_new(lp->gid, offset_ts, lp);
    model_net_wrap_msg *m_wrap = tw_event_data(e);
679 680
    msg_set_header(model_net_base_magic, MN_BASE_SCHED_NEXT, lp->gid,
            &m_wrap->h);
681
    m_wrap->msg.m_base.is_from_remote = is_recv_queue;
682 683 684
    tw_event_send(e);
}

685 686 687 688 689 690 691 692 693 694 695 696
void * model_net_method_get_edata(int net_id, void *msg){
    return (char*)msg + sizeof(model_net_wrap_msg) - msg_offsets[net_id];
}

/*
 * Local variables:
 *  c-indent-level: 4
 *  c-basic-offset: 4
 * End:
 *
 * vim: ft=c ts=8 sts=4 sw=4 expandtab
 */