margo.c 29.2 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11

/*
 * (C) 2015 The University of Chicago
 * 
 * See COPYRIGHT in top-level directory.
 */

#include <assert.h>
#include <unistd.h>
#include <errno.h>
#include <abt.h>
12
#include <abt-snoozer.h>
13
#include <time.h>
Philip Carns's avatar
Philip Carns committed
14
#include <math.h>
15 16

#include "margo.h"
17
#include "margo-timer.h"
Philip Carns's avatar
Philip Carns committed
18
#include "utlist.h"
19
#include "uthash.h"
20

21 22
#define MERCURY_PROGRESS_TIMEOUT_UB 100 /* 100 milliseconds */

23 24 25 26 27 28 29 30 31 32 33 34 35
struct mplex_key
{
    hg_id_t id;
    uint32_t mplex_id;
};

struct mplex_element
{
    struct mplex_key key;
    ABT_pool pool;
    UT_hash_handle hh;
};

Philip Carns's avatar
Philip Carns committed
36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51
struct diag_data
{
    double min;
    double max;
    double cumulative;
    int count;
};

#define __DIAG_UPDATE(__data, __time)\
do {\
    __data.count++; \
    __data.cumulative += (__time); \
    if((__time) > __data.max) __data.max = (__time); \
    if((__time) < __data.min) __data.min = (__time); \
} while(0)

52 53
struct margo_instance
{
54
    /* provided by caller */
55 56
    hg_context_t *hg_context;
    hg_class_t *hg_class;
57 58 59
    ABT_pool handler_pool;
    ABT_pool progress_pool;

60
    /* internal to margo for this particular instance */
61
    int margo_init;
62 63
    ABT_thread hg_progress_tid;
    int hg_progress_shutdown_flag;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
64
    ABT_xstream progress_xstream;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
65 66 67
    int owns_progress_pool;
    ABT_xstream *rpc_xstreams;
    int num_handler_pool_threads;
68 69 70

    /* control logic for callers waiting on margo to be finalized */
    int finalize_flag;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
71
    int refcount;
72 73 74
    ABT_mutex finalize_mutex;
    ABT_cond finalize_cond;

75 76
    /* hash table to track multiplexed rpcs registered with margo */
    struct mplex_element *mplex_table;
Philip Carns's avatar
Philip Carns committed
77 78 79 80 81 82 83 84 85 86 87 88

    /* optional diagnostics data tracking */
    /* NOTE: technically the following fields are subject to races if they
     * are updated from more than one thread at a time.  We will be careful
     * to only update the counters from the progress_fn,
     * which will serialize access.
     */
    int diag_enabled;
    struct diag_data diag_trigger_elapsed;
    struct diag_data diag_progress_elapsed_zero_timeout;
    struct diag_data diag_progress_elapsed_nonzero_timeout;
    struct diag_data diag_progress_timeout_value;
89 90
};

91 92 93 94 95 96
struct margo_cb_arg
{
    ABT_eventual *eventual;
    margo_instance_id mid;
};

97 98 99 100 101 102
struct margo_rpc_data
{
	margo_instance_id mid;
	void* user_data;
	void (*user_free_callback)(void *);
};
103

104
static void hg_progress_fn(void* foo);
105
static void margo_rpc_data_free(void* ptr);
106

107
margo_instance_id margo_init(const char *addr_str, int mode,
108
    int use_progress_thread, int rpc_thread_count)
109
{
Jonathan Jenkins's avatar
Jonathan Jenkins committed
110 111 112 113 114
    ABT_xstream progress_xstream = ABT_XSTREAM_NULL;
    ABT_pool progress_pool = ABT_POOL_NULL;
    ABT_xstream *rpc_xstreams = NULL;
    ABT_xstream rpc_xstream = ABT_XSTREAM_NULL;
    ABT_pool rpc_pool = ABT_POOL_NULL;
115 116
    hg_class_t *hg_class = NULL;
    hg_context_t *hg_context = NULL;
117
    int listen_flag = (mode == MARGO_CLIENT_MODE) ? HG_FALSE : HG_TRUE;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
118
    int i;
119 120 121
    int ret;
    struct margo_instance *mid = MARGO_INSTANCE_NULL;

Shane Snyder's avatar
Shane Snyder committed
122
    if(mode != MARGO_CLIENT_MODE && mode != MARGO_SERVER_MODE) goto err;
123

124 125 126
    ret = ABT_init(0, NULL); /* XXX: argc/argv not currently used by ABT ... */
    if(ret != 0) goto err;

127
    /* set caller (self) ES to idle without polling */
128 129
    ret = ABT_snoozer_xstream_self_set();
    if(ret != 0) goto err;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
130 131 132 133 134 135 136 137 138 139 140 141 142 143

    if (use_progress_thread)
    {
        ret = ABT_snoozer_xstream_create(1, &progress_pool, &progress_xstream);
        if (ret != ABT_SUCCESS) goto err;
    }
    else
    {
        ret = ABT_xstream_self(&progress_xstream);
        if (ret != ABT_SUCCESS) goto err;
        ret = ABT_xstream_get_main_pools(progress_xstream, 1, &progress_pool);
        if (ret != ABT_SUCCESS) goto err;
    }

144
    if (mode == MARGO_SERVER_MODE)
Jonathan Jenkins's avatar
Jonathan Jenkins committed
145
    {
146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164
        if (rpc_thread_count > 0)
        {
            rpc_xstreams = malloc(rpc_thread_count * sizeof(*rpc_xstreams));
            if (rpc_xstreams == NULL) goto err;
            ret = ABT_snoozer_xstream_create(rpc_thread_count, &rpc_pool,
                    rpc_xstreams);
            if (ret != ABT_SUCCESS) goto err;
        }
        else if (rpc_thread_count == 0)
        {
            ret = ABT_xstream_self(&rpc_xstream);
            if (ret != ABT_SUCCESS) goto err;
            ret = ABT_xstream_get_main_pools(rpc_xstream, 1, &rpc_pool);
            if (ret != ABT_SUCCESS) goto err;
        }
        else
        {
            rpc_pool = progress_pool;
        }
Jonathan Jenkins's avatar
Jonathan Jenkins committed
165 166
    }

167 168 169 170 171 172
    hg_class = HG_Init(addr_str, listen_flag);
    if(!hg_class) goto err;

    hg_context = HG_Context_create(hg_class);
    if(!hg_context) goto err;

Jonathan Jenkins's avatar
Jonathan Jenkins committed
173 174 175
    mid = margo_init_pool(progress_pool, rpc_pool, hg_context);
    if (mid == MARGO_INSTANCE_NULL) goto err;

176
    mid->margo_init = 1;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
177 178 179
    mid->owns_progress_pool = use_progress_thread;
    mid->progress_xstream = progress_xstream;
    mid->num_handler_pool_threads = rpc_thread_count < 0 ? 0 : rpc_thread_count;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
180 181 182 183
    mid->rpc_xstreams = rpc_xstreams;
    return mid;

err:
184 185 186 187 188 189 190
    if(mid)
    {
        margo_timer_instance_finalize(mid);
        ABT_mutex_free(&mid->finalize_mutex);
        ABT_cond_free(&mid->finalize_cond);
        free(mid);
    }
Jonathan Jenkins's avatar
Jonathan Jenkins committed
191 192 193 194 195 196 197 198 199 200 201 202 203 204
    if (use_progress_thread && progress_xstream != ABT_XSTREAM_NULL)
    {
        ABT_xstream_join(progress_xstream);
        ABT_xstream_free(&progress_xstream);
    }
    if (rpc_thread_count > 0 && rpc_xstreams != NULL)
    {
        for (i = 0; i < rpc_thread_count; i++)
        {
            ABT_xstream_join(rpc_xstreams[i]);
            ABT_xstream_free(&rpc_xstreams[i]);
        }
        free(rpc_xstreams);
    }
205 206 207 208 209
    if(hg_context)
        HG_Context_destroy(hg_context);
    if(hg_class)
        HG_Finalize(hg_class);
    ABT_finalize();
Jonathan Jenkins's avatar
Jonathan Jenkins committed
210 211 212 213
    return MARGO_INSTANCE_NULL;
}

margo_instance_id margo_init_pool(ABT_pool progress_pool, ABT_pool handler_pool,
Jonathan Jenkins's avatar
Jonathan Jenkins committed
214
    hg_context_t *hg_context)
215 216
{
    int ret;
217 218 219
    struct margo_instance *mid;

    mid = malloc(sizeof(*mid));
220
    if(!mid) goto err;
221
    memset(mid, 0, sizeof(*mid));
222

223 224 225
    ABT_mutex_create(&mid->finalize_mutex);
    ABT_cond_create(&mid->finalize_cond);

226 227
    mid->progress_pool = progress_pool;
    mid->handler_pool = handler_pool;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
228
    mid->hg_class = HG_Context_get_class(hg_context);
229
    mid->hg_context = hg_context;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
230
    mid->refcount = 1;
231

232
    ret = margo_timer_instance_init(mid);
233
    if(ret != 0) goto err;
234

235
    ret = ABT_thread_create(mid->progress_pool, hg_progress_fn, mid, 
236
        ABT_THREAD_ATTR_NULL, &mid->hg_progress_tid);
237 238 239 240
    if(ret != 0) goto err;

err:
    if(mid)
241
    {
242 243 244
        margo_timer_instance_finalize(mid);
        ABT_mutex_free(&mid->finalize_mutex);
        ABT_cond_free(&mid->finalize_cond);
245
        free(mid);
246
    }
247
    return MARGO_INSTANCE_NULL;
248 249
}

Jonathan Jenkins's avatar
Jonathan Jenkins committed
250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274
static void margo_cleanup(margo_instance_id mid)
{
    int i;

    margo_timer_instance_finalize(mid);

    ABT_mutex_free(&mid->finalize_mutex);
    ABT_cond_free(&mid->finalize_cond);

    if (mid->owns_progress_pool)
    {
        ABT_xstream_join(mid->progress_xstream);
        ABT_xstream_free(&mid->progress_xstream);
    }

    if (mid->num_handler_pool_threads > 0)
    {
        for (i = 0; i < mid->num_handler_pool_threads; i++)
        {
            ABT_xstream_join(mid->rpc_xstreams[i]);
            ABT_xstream_free(&mid->rpc_xstreams[i]);
        }
        free(mid->rpc_xstreams);
    }

275 276 277 278 279 280 281 282 283
    if (mid->margo_init)
    {
        if (mid->hg_context)
            HG_Context_destroy(mid->hg_context);
        if (mid->hg_class)
            HG_Finalize(mid->hg_class);
        ABT_finalize();
    }

Jonathan Jenkins's avatar
Jonathan Jenkins committed
284 285 286
    free(mid);
}

287
void margo_finalize(margo_instance_id mid)
288
{
Jonathan Jenkins's avatar
Jonathan Jenkins committed
289
    int do_cleanup;
290

291
    /* tell progress thread to wrap things up */
292
    mid->hg_progress_shutdown_flag = 1;
293 294

    /* wait for it to shutdown cleanly */
295 296
    ABT_thread_join(mid->hg_progress_tid);
    ABT_thread_free(&mid->hg_progress_tid);
297

298 299 300 301
    ABT_mutex_lock(mid->finalize_mutex);
    mid->finalize_flag = 1;
    ABT_cond_broadcast(mid->finalize_cond);

Jonathan Jenkins's avatar
Jonathan Jenkins committed
302 303
    mid->refcount--;
    do_cleanup = mid->refcount == 0;
304

Jonathan Jenkins's avatar
Jonathan Jenkins committed
305 306 307 308 309 310 311
    ABT_mutex_unlock(mid->finalize_mutex);

    /* if there was noone waiting on the finalize at the time of the finalize
     * broadcast, then we're safe to clean up. Otherwise, let the finalizer do
     * it */
    if (do_cleanup)
        margo_cleanup(mid);
312 313 314 315 316 317

    return;
}

void margo_wait_for_finalize(margo_instance_id mid)
{
Jonathan Jenkins's avatar
Jonathan Jenkins committed
318
    int do_cleanup;
319 320 321

    ABT_mutex_lock(mid->finalize_mutex);

Jonathan Jenkins's avatar
Jonathan Jenkins committed
322
        mid->refcount++;
323 324 325 326
            
        while(!mid->finalize_flag)
            ABT_cond_wait(mid->finalize_cond, mid->finalize_mutex);

Jonathan Jenkins's avatar
Jonathan Jenkins committed
327 328 329
        mid->refcount--;
        do_cleanup = mid->refcount == 0;

330
    ABT_mutex_unlock(mid->finalize_mutex);
Jonathan Jenkins's avatar
Jonathan Jenkins committed
331 332 333 334

    if (do_cleanup)
        margo_cleanup(mid);

335 336 337
    return;
}

338 339
hg_id_t margo_register_name(margo_instance_id mid, const char *func_name,
    hg_proc_cb_t in_proc_cb, hg_proc_cb_t out_proc_cb, hg_rpc_cb_t rpc_cb)
340
{
341 342 343
	struct margo_rpc_data* margo_data;
    hg_return_t hret;
    hg_id_t id;
344

345 346 347
    id = HG_Register_name(mid->hg_class, func_name, in_proc_cb, out_proc_cb, rpc_cb);
    if(id <= 0)
        return(0);
348

349 350 351 352 353 354 355 356 357 358 359 360
	/* register the margo data with the RPC */
    margo_data = (struct margo_rpc_data*)HG_Registered_data(mid->hg_class, id);
    if(!margo_data)
    {
        margo_data = (struct margo_rpc_data*)malloc(sizeof(struct margo_rpc_data));
        if(!margo_data)
            return(0);
        margo_data->mid = mid;
        margo_data->user_data = NULL;
        margo_data->user_free_callback = NULL;
        hret = HG_Register_data(mid->hg_class, id, margo_data, margo_rpc_data_free);
        if(hret != HG_SUCCESS)
361
        {
362 363
            free(margo_data);
            return(0);
364
        }
365 366
    }

367
	return(id);
368 369
}

370 371 372
hg_id_t margo_register_name_mplex(margo_instance_id mid, const char *func_name,
    hg_proc_cb_t in_proc_cb, hg_proc_cb_t out_proc_cb, hg_rpc_cb_t rpc_cb,
    uint32_t mplex_id, ABT_pool pool)
373
{
374 375 376
    struct mplex_key key;
    struct mplex_element *element;
    hg_id_t id;
377

378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402
    id = margo_register_name(mid, func_name, in_proc_cb, out_proc_cb, rpc_cb);
    if(id <= 0)
        return(0);

    /* nothing to do, we'll let the handler pool take this directly */
    if(mplex_id == MARGO_DEFAULT_MPLEX_ID)
        return(id);

    memset(&key, 0, sizeof(key));
    key.id = id;
    key.mplex_id = mplex_id;

    HASH_FIND(hh, mid->mplex_table, &key, sizeof(key), element);
    if(element)
        return(id);

    element = malloc(sizeof(*element));
    if(!element)
        return(0);
    element->key = key;
    element->pool = pool;

    HASH_ADD(hh, mid->mplex_table, key, sizeof(key), element);

    return(id);
403 404
}

405 406
hg_return_t margo_registered_name(margo_instance_id mid, const char *func_name,
    hg_id_t *id, hg_bool_t *flag)
407
{
408
    return(HG_Registered_name(mid->hg_class, func_name, id, flag));
409 410
}

411 412 413 414 415 416 417
hg_return_t margo_register_data(
    margo_instance_id mid,
    hg_id_t id,
    void *data,
    void (*free_callback)(void *)) 
{
	struct margo_rpc_data* margo_data 
418
		= (struct margo_rpc_data*) HG_Registered_data(mid->hg_class, id);
419 420 421 422 423 424 425 426 427 428 429 430 431 432
	if(!margo_data) return HG_OTHER_ERROR;
	margo_data->user_data = data;
	margo_data->user_free_callback = free_callback;
	return HG_SUCCESS;
}

void* margo_registered_data(margo_instance_id mid, hg_id_t id)
{
	struct margo_rpc_data* data
		= (struct margo_rpc_data*) HG_Registered_data(margo_get_class(mid), id);
	if(!data) return NULL;
	else return data->user_data;
}

433 434 435 436
hg_return_t margo_registered_disable_response(
    margo_instance_id mid,
    hg_id_t id,
    int disable_flag)
437
{
438
    return(HG_Registered_disable_response(mid->hg_class, id, disable_flag));
439
}
440

441
struct lookup_cb_evt
442
{
443
    hg_return_t hret;
444 445 446 447 448 449
    hg_addr_t addr;
};

static hg_return_t margo_addr_lookup_cb(const struct hg_cb_info *info)
{
    struct lookup_cb_evt evt;
450
    evt.hret = info->ret;
451
    evt.addr = info->info.lookup.addr;
452
    struct margo_cb_arg* arg = info->arg;
453 454

    /* propagate return code out through eventual */
455 456
    ABT_eventual_set(*(arg->eventual), &evt, sizeof(evt));

457 458 459
    return(HG_SUCCESS);
}

460 461 462 463
hg_return_t margo_addr_lookup(
    margo_instance_id mid,
    const char   *name,
    hg_addr_t    *addr)
464
{
465
    hg_return_t hret;
466 467 468 469
    struct lookup_cb_evt *evt;
    ABT_eventual eventual;
    int ret;
    struct margo_cb_arg arg;
470

471 472 473 474 475 476 477 478
    ret = ABT_eventual_create(sizeof(*evt), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

    arg.eventual = &eventual;
    arg.mid = mid;
479

480
    hret = HG_Addr_lookup(mid->hg_context, margo_addr_lookup_cb,
481
        &arg, name, HG_OP_ID_IGNORE);
482
    if(hret == HG_SUCCESS)
483 484 485
    {
        ABT_eventual_wait(eventual, (void**)&evt);
        *addr = evt->addr;
486
        hret = evt->hret;
487 488 489 490
    }

    ABT_eventual_free(&eventual);

491
    return(hret);
492 493 494 495 496
}

hg_return_t margo_addr_free(
    margo_instance_id mid,
    hg_addr_t addr)
497
{
498 499
    return(HG_Addr_free(mid->hg_class, addr));
}
500

501 502 503 504 505
hg_return_t margo_addr_self(
    margo_instance_id mid,
    hg_addr_t *addr)
{
    return(HG_Addr_self(mid->hg_class, addr));
506 507
}

508 509 510 511 512 513 514 515 516
hg_return_t margo_addr_dup(
    margo_instance_id mid,
    hg_addr_t addr,
    hg_addr_t *new_addr)
{
    return(HG_Addr_dup(mid->hg_class, addr, new_addr));
}

hg_return_t margo_addr_to_string(
517
    margo_instance_id mid,
518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552
    char *buf,
    hg_size_t *buf_size,
    hg_addr_t addr)
{
    return(HG_Addr_to_string(mid->hg_class, buf, buf_size, addr));
}

hg_return_t margo_create(margo_instance_id mid, hg_addr_t addr,
    hg_id_t id, hg_handle_t *handle)
{
    /* TODO: handle caching logic? */

    return(HG_Create(mid->hg_context, addr, id, handle));
}

hg_return_t margo_destroy(hg_handle_t handle)
{
    /* TODO handle caching logic? */

    return(HG_Destroy(handle));
}

hg_return_t margo_ref_incr(
    hg_handle_t handle)
{
    return(HG_Ref_incr(handle));
}

const struct hg_info *margo_get_info(
    hg_handle_t handle)
{
    return(HG_Get_info(handle));
}

hg_return_t margo_get_input(
553
    hg_handle_t handle,
554
    void *in_struct)
555
{
556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596
    return(HG_Get_input(handle, in_struct));
}

hg_return_t margo_free_input(
        hg_handle_t handle,
        void *in_struct)
{
    return(HG_Free_input(handle, in_struct));
}

hg_return_t margo_get_output(
    hg_handle_t handle,
    void *out_struct)
{
    return(HG_Get_output(handle, out_struct));
}

hg_return_t margo_free_output(
    hg_handle_t handle,
    void *out_struct)
{
    return(HG_Free_output(handle, out_struct));
}

static hg_return_t margo_cb(const struct hg_cb_info *info)
{
    hg_return_t hret = info->ret;
    struct margo_cb_arg* arg = info->arg;

    /* propagate return code out through eventual */
    ABT_eventual_set(*(arg->eventual), &hret, sizeof(hret));
    
    return(HG_SUCCESS);
}

hg_return_t margo_forward(
    margo_instance_id mid,
    hg_handle_t handle,
    void *in_struct)
{
    hg_return_t hret = HG_TIMEOUT;
597
    ABT_eventual eventual;
598
    int ret;
599
    hg_return_t* waited_hret;
600
    struct margo_cb_arg arg;
601 602 603 604 605 606 607

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

608 609
    arg.eventual = &eventual;
    arg.mid = mid;
610

611
    hret = HG_Forward(handle, margo_cb, &arg, in_struct);
612
    if(hret == HG_SUCCESS)
613 614 615 616 617 618 619 620
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

    ABT_eventual_free(&eventual);

    return(hret);
621 622
}

623 624 625 626
typedef struct
{
    hg_handle_t handle;
} margo_forward_timeout_cb_dat;
627

628 629 630 631 632 633 634 635 636 637 638
static void margo_forward_timeout_cb(void *arg)
{
    margo_forward_timeout_cb_dat *timeout_cb_dat =
        (margo_forward_timeout_cb_dat *)arg;

    /* cancel the Mercury op if the forward timed out */
    HG_Cancel(timeout_cb_dat->handle);
    return;
}

hg_return_t margo_forward_timed(
639
    margo_instance_id mid,
640
    hg_handle_t handle,
641 642
    void *in_struct,
    double timeout_ms)
643 644
{
    int ret;
645 646
    hg_return_t hret;
    ABT_eventual eventual;
647
    hg_return_t* waited_hret;
648 649
    margo_timer_t forward_timer;
    margo_forward_timeout_cb_dat timeout_cb_dat;
650
    struct margo_cb_arg arg;
651 652 653 654 655 656 657

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

658 659 660 661 662
    /* set a timer object to expire when this forward times out */
    timeout_cb_dat.handle = handle;
    margo_timer_init(mid, &forward_timer, margo_forward_timeout_cb,
        &timeout_cb_dat, timeout_ms);

663 664
    arg.eventual = &eventual;
    arg.mid = mid;
665

666
    hret = HG_Forward(handle, margo_cb, &arg, in_struct);
667
    if(hret == HG_SUCCESS)
Jonathan Jenkins's avatar
Jonathan Jenkins committed
668 669 670 671 672
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

673 674 675 676 677 678 679 680
    /* convert HG_CANCELED to HG_TIMEOUT to indicate op timed out */
    if(hret == HG_CANCELED)
        hret = HG_TIMEOUT;

    /* remove timer if it is still in place (i.e., not timed out) */
    if(hret != HG_TIMEOUT)
        margo_timer_destroy(mid, &forward_timer);

Jonathan Jenkins's avatar
Jonathan Jenkins committed
681 682 683 684 685 686 687 688 689 690 691 692 693 694
    ABT_eventual_free(&eventual);

    return(hret);
}

hg_return_t margo_respond(
    margo_instance_id mid,
    hg_handle_t handle,
    void *out_struct)
{
    hg_return_t hret = HG_TIMEOUT;
    ABT_eventual eventual;
    int ret;
    hg_return_t* waited_hret;
695
    struct margo_cb_arg arg;
Jonathan Jenkins's avatar
Jonathan Jenkins committed
696 697 698 699 700 701 702

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);
    }

703 704
    arg.eventual = &eventual;
    arg.mid = mid;
705

706
    hret = HG_Respond(handle, margo_cb, &arg, out_struct);
707
    if(hret == HG_SUCCESS)
708 709 710 711 712 713 714 715 716 717
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

    ABT_eventual_free(&eventual);

    return(hret);
}

718 719 720 721 722
hg_return_t margo_cancel(
    hg_handle_t handle)
{
    return(HG_Cancel(handle));
}
723

724 725 726 727 728 729 730
hg_return_t margo_bulk_create(
    margo_instance_id mid,
    hg_uint32_t count,
    void **buf_ptrs,
    const hg_size_t *buf_sizes,
    hg_uint8_t flags,
    hg_bulk_t *handle)
731
{
732
    /* XXX: handle caching logic? */
733

734 735 736
    return(HG_Bulk_create(mid->hg_class, count,
        buf_ptrs, buf_sizes, flags, handle));
}
737

738 739 740 741 742 743
hg_return_t margo_bulk_free(
    hg_bulk_t handle)
{
    /* XXX: handle caching logic? */

    return(HG_Bulk_free(handle));
744 745
}

746 747
hg_return_t margo_bulk_ref_incr(
    hg_bulk_t handle)
Philip Carns's avatar
Philip Carns committed
748
{
749 750
    return(HG_Bulk_ref_incr(handle));
}
Philip Carns's avatar
Philip Carns committed
751

752 753 754 755 756 757 758 759 760
hg_return_t margo_bulk_access(
    hg_bulk_t handle,
    hg_size_t offset,
    hg_size_t size,
    hg_uint8_t flags,
    hg_uint32_t max_count,
    void **buf_ptrs,
    hg_size_t *buf_sizes,
    hg_uint32_t *actual_count)
Philip Carns's avatar
Philip Carns committed
761
{
762 763 764
    return(HG_Bulk_access(handle, offset, size, flags, max_count,
        buf_ptrs, buf_sizes, actual_count));
}
Philip Carns's avatar
Philip Carns committed
765

766 767 768 769 770
hg_size_t margo_bulk_get_size(
    hg_bulk_t handle)
{
    return(HG_Bulk_get_size(handle));
}
771

772 773 774 775
hg_uint32_t margo_bulk_get_segment_count(
    hg_bulk_t handle)
{
    return(HG_Bulk_get_segment_count(handle));
Philip Carns's avatar
Philip Carns committed
776 777
}

778 779 780 781 782 783
hg_size_t margo_bulk_get_serialize_size(
    hg_bulk_t handle,
    hg_bool_t request_eager)
{
    return(HG_Bulk_get_serialize_size(handle, request_eager));
}
Philip Carns's avatar
Philip Carns committed
784

785 786 787 788 789
hg_return_t margo_bulk_serialize(
    void *buf,
    hg_size_t buf_size,
    hg_bool_t request_eager,
    hg_bulk_t handle)
790
{
791 792
    return(HG_Bulk_serialize(buf, buf_size, request_eager, handle));
}
793

794 795 796 797 798 799 800 801
hg_return_t margo_bulk_deserialize(
    margo_instance_id mid,
    hg_bulk_t *handle,
    const void *buf,
    hg_size_t buf_size)
{
    return(HG_Bulk_deserialize(mid->hg_class, handle, buf, buf_size));
}
802

803 804 805 806
static hg_return_t margo_bulk_transfer_cb(const struct hg_cb_info *info)
{
    hg_return_t hret = info->ret;
    struct margo_cb_arg* arg = info->arg;
807

808 809 810 811
    /* propagate return code out through eventual */
    ABT_eventual_set(*(arg->eventual), &hret, sizeof(hret));
    
    return(HG_SUCCESS);
812 813
}

814
hg_return_t margo_bulk_transfer(
815
    margo_instance_id mid,
816
    hg_bulk_op_t op,
817
    hg_addr_t origin_addr,
818 819 820 821
    hg_bulk_t origin_handle,
    size_t origin_offset,
    hg_bulk_t local_handle,
    size_t local_offset,
822 823
    size_t size,
    hg_op_id_t *op_id)
824 825 826 827 828
{
    hg_return_t hret = HG_TIMEOUT;
    hg_return_t *waited_hret;
    ABT_eventual eventual;
    int ret;
829
    struct margo_cb_arg arg;
830 831 832 833 834 835 836

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

837 838
    arg.eventual = &eventual;
    arg.mid = mid;
839

Jonathan Jenkins's avatar
Jonathan Jenkins committed
840
    hret = HG_Bulk_transfer(mid->hg_context, margo_bulk_transfer_cb,
841
        &arg, op, origin_addr, origin_handle, origin_offset, local_handle,
842
        local_offset, size, op_id);
843
    if(hret == HG_SUCCESS)
844 845 846 847 848 849 850 851 852 853
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

    ABT_eventual_free(&eventual);

    return(hret);
}

854 855 856 857 858 859
hg_return_t margo_bulk_cancel(
    hg_op_id_t op_id)
{
    return(HG_Bulk_cancel(op_id));
}

860 861 862 863
typedef struct
{
    ABT_mutex mutex;
    ABT_cond cond;
Shane Snyder's avatar
Shane Snyder committed
864
    char is_asleep;
865 866 867 868 869 870 871 872 873
} margo_thread_sleep_cb_dat;

static void margo_thread_sleep_cb(void *arg)
{
    margo_thread_sleep_cb_dat *sleep_cb_dat =
        (margo_thread_sleep_cb_dat *)arg;

    /* wake up the sleeping thread */
    ABT_mutex_lock(sleep_cb_dat->mutex);
874
    sleep_cb_dat->is_asleep = 0;
875 876 877 878 879 880 881
    ABT_cond_signal(sleep_cb_dat->cond);
    ABT_mutex_unlock(sleep_cb_dat->mutex);

    return;
}

void margo_thread_sleep(
882
    margo_instance_id mid,
883 884 885 886 887 888 889 890
    double timeout_ms)
{
    margo_timer_t sleep_timer;
    margo_thread_sleep_cb_dat sleep_cb_dat;

    /* set data needed for sleep callback */
    ABT_mutex_create(&(sleep_cb_dat.mutex));
    ABT_cond_create(&(sleep_cb_dat.cond));
891
    sleep_cb_dat.is_asleep = 1;
892 893

    /* initialize the sleep timer */
894
    margo_timer_init(mid, &sleep_timer, margo_thread_sleep_cb,
895 896 897 898
        &sleep_cb_dat, timeout_ms);

    /* yield thread for specified timeout */
    ABT_mutex_lock(sleep_cb_dat.mutex);
899 900
    while(sleep_cb_dat.is_asleep)
        ABT_cond_wait(sleep_cb_dat.cond, sleep_cb_dat.mutex);
901 902
    ABT_mutex_unlock(sleep_cb_dat.mutex);

903 904 905 906
    /* clean up */
    ABT_mutex_free(&sleep_cb_dat.mutex);
    ABT_cond_free(&sleep_cb_dat.cond);

907 908 909
    return;
}

910
ABT_pool* margo_get_handler_pool(margo_instance_id mid)
911
{
912 913
    return(&mid->handler_pool);
}
914

915 916 917 918
hg_context_t* margo_get_context(margo_instance_id mid)
{
    return(mid->hg_context);
}
919

920 921 922
hg_class_t* margo_get_class(margo_instance_id mid)
{
    return(mid->hg_class);
923
}
Philip Carns's avatar
Philip Carns committed
924

925
margo_instance_id margo_hg_handle_get_instance(hg_handle_t h)
926
{
927 928 929 930 931 932
	const struct hg_info* info = HG_Get_info(h);
	if(!info) return MARGO_INSTANCE_NULL;
	struct margo_rpc_data* data = 
		(struct margo_rpc_data*) HG_Registered_data(info->hg_class, info->id);
	if(!data) return MARGO_INSTANCE_NULL;
	return data->mid;
933 934
}

935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953
int margo_lookup_mplex(margo_instance_id mid, hg_id_t id, uint32_t mplex_id, ABT_pool *pool)
{
    struct mplex_key key;
    struct mplex_element *element;

    if(!mplex_id)
    {
        *pool = mid->handler_pool;
        return(0);
    }

    memset(&key, 0, sizeof(key));
    key.id = id;
    key.mplex_id = mplex_id;

    HASH_FIND(hh, mid->mplex_table, &key, sizeof(key), element);
    if(!element)
        return(-1);

Philip Carns's avatar
Philip Carns committed
954 955
    assert(element->key.id == id && element->key.mplex_id == mplex_id);

956 957 958 959 960
    *pool = element->pool;

    return(0);
}

961
static void margo_rpc_data_free(void* ptr)
Philip Carns's avatar
Philip Carns committed
962
{
963 964 965 966 967 968
	struct margo_rpc_data* data = (struct margo_rpc_data*) ptr;
	if(data->user_data && data->user_free_callback) {
		data->user_free_callback(data->user_data);
	}
	free(ptr);
}
969

970 971 972 973 974 975 976 977 978 979
/* dedicated thread function to drive Mercury progress */
static void hg_progress_fn(void* foo)
{
    int ret;
    unsigned int actual_count;
    struct margo_instance *mid = (struct margo_instance *)foo;
    size_t size;
    unsigned int hg_progress_timeout = MERCURY_PROGRESS_TIMEOUT_UB;
    double next_timer_exp;
    int trigger_happened;
980 981
    double tm1, tm2;
    int diag_enabled = 0;
982

983 984 985 986
    while(!mid->hg_progress_shutdown_flag)
    {
        trigger_happened = 0;
        do {
987 988 989 990
            /* save value of instance diag variable, in case it is modified
             * while we are in loop 
             */
            diag_enabled = mid->diag_enabled;
991

992
            if(diag_enabled) tm1 = ABT_get_wtime();
993
            ret = HG_Trigger(mid->hg_context, 0, 1, &actual_count);
994 995 996 997 998
            if(diag_enabled)
            {
                tm2 = ABT_get_wtime();
                __DIAG_UPDATE(mid->diag_trigger_elapsed, (tm2-tm1));
            }
999

1000 1001 1002
            if(ret == HG_SUCCESS && actual_count > 0)
                trigger_happened = 1;
        } while((ret == HG_SUCCESS) && actual_count && !mid->hg_progress_shutdown_flag);
1003

1004 1005
        if(trigger_happened)
            ABT_thread_yield();
1006

1007
        ABT_pool_get_size(mid->progress_pool, &size);
1008
        /* Are there any other threads executing in this pool that are *not*
1009 1010 1011 1012
         * blocked ?  If so then, we can't sleep here or else those threads 
         * will not get a chance to execute.
         * TODO: check is ABT_pool_get_size returns the number of ULT/tasks
         * that can be executed including this one, or not including this one.
1013
         */
1014
        if(size > 0)
1015 1016 1017 1018 1019 1020 1021 1022
        {
            /* TODO: this is being executed more than is necessary (i.e.
             * in cases where there are other legitimate ULTs eligible
             * for execution that are not blocking on any events, Margo
             * or otherwise). Maybe we need an abt scheduling tweak here
             * to make sure that this ULT is the lowest priority in that
             * scenario.
             */
1023
            if(diag_enabled) tm1 = ABT_get_wtime();
1024
            ret = HG_Progress(mid->hg_context, 0);
1025 1026 1027 1028 1029 1030
            if(diag_enabled)
            {
                tm2 = ABT_get_wtime();
                __DIAG_UPDATE(mid->diag_progress_elapsed_zero_timeout, (tm2-tm1));
                __DIAG_UPDATE(mid->diag_progress_timeout_value, 0);
            }
1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067
            if(ret == HG_SUCCESS)
            {
                /* Mercury completed something; loop around to trigger
                 * callbacks 
                 */
            }
            else if(ret == HG_TIMEOUT)
            {
                /* No completion; yield here to allow other ULTs to run */
                ABT_thread_yield();
            }
            else
            {
                /* TODO: error handling */
                fprintf(stderr, "WARNING: unexpected return code (%d) from HG_Progress()\n", ret);
            }
        }
        else
        {
            hg_progress_timeout = MERCURY_PROGRESS_TIMEOUT_UB;
            ret = margo_timer_get_next_expiration(mid, &next_timer_exp);
            if(ret == 0)
            {
                /* there is a queued timer, don't block long enough
                 * to keep this timer waiting
                 */
                if(next_timer_exp >= 0.0)
                {
                    next_timer_exp *= 1000; /* convert to milliseconds */
                    if(next_timer_exp < MERCURY_PROGRESS_TIMEOUT_UB)
                        hg_progress_timeout = (unsigned int)next_timer_exp;
                }
                else
                {
                    hg_progress_timeout = 0;
                }
            }
1068
            if(diag_enabled) tm1 = ABT_get_wtime();
1069
            ret = HG_Progress(mid->hg_context, hg_progress_timeout);
1070 1071 1072 1073 1074 1075 1076 1077 1078 1079
            if(diag_enabled)
            {
                tm2 = ABT_get_wtime();
                if(hg_progress_timeout == 0)
                    __DIAG_UPDATE(mid->diag_progress_elapsed_zero_timeout, (tm2-tm1));
                else
                    __DIAG_UPDATE(mid->diag_progress_elapsed_nonzero_timeout, (tm2-tm1));
                    
                __DIAG_UPDATE(mid->diag_progress_timeout_value, hg_progress_timeout);
            }
1080 1081 1082 1083 1084 1085
            if(ret != HG_SUCCESS && ret != HG_TIMEOUT)
            {
                /* TODO: error handling */
                fprintf(stderr, "WARNING: unexpected return code (%d) from HG_Progress()\n", ret);
            }
        }
1086

1087 1088 1089
        /* check for any expired timers */
        margo_check_timers(mid);
    }
1090

1091
    return;
Philip Carns's avatar
Philip Carns committed
1092
}
Philip Carns's avatar
Philip Carns committed
1093 1094 1095 1096 1097 1098 1099


void margo_diag_start(margo_instance_id mid)
{
    mid->diag_enabled = 1;
}

1100
static void print_diag_data(FILE *file, const char* name, const char *description, struct diag_data *data)
Philip Carns's avatar
Philip Carns committed
1101
{
1102 1103 1104 1105 1106 1107 1108 1109
    double avg;

    fprintf(file, "# %s\n", description);
    if(data->count != 0)
        avg = data->cumulative/data->count;
    else
        avg = 0;
    fprintf(file, "%s\t%.9f\t%.9f\t%.9f\t%d\n", name, avg, data->min, data->max, data->count);
Philip Carns's avatar
Philip Carns committed
1110 1111 1112
    return;
}

1113
void margo_diag_dump(margo_instance_id mid, const char* file, int uniquify)
Philip Carns's avatar
Philip Carns committed
1114 1115 1116
{
    FILE *outfile;
    time_t ltime;
1117
    char revised_file_name[256] = {0};
Philip Carns's avatar
Philip Carns committed
1118 1119 1120

    assert(mid->diag_enabled);

1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135
    if(uniquify)
    {
        char hostname[128] = {0};
        int pid;

        gethostname(hostname, 128);
        pid = getpid();

        sprintf(revised_file_name, "%s-%s-%d", file, hostname, pid);
    }
    else
    {
        sprintf(revised_file_name, "%s", file);
    }

Philip Carns's avatar
Philip Carns committed
1136 1137 1138 1139 1140 1141
    if(strcmp("-", file) == 0)
    {
        outfile = stdout;
    }
    else
    {
1142
        outfile = fopen(revised_file_name, "a");
Philip Carns's avatar
Philip Carns committed
1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158
        if(!outfile)
        {
            perror("fopen");
            return;
        }
    }

    /* TODO: retrieve self addr and include in output */
    /* TODO: support pattern substitution in file name to create unique
     * output files per process
     */

    time(&ltime);
    fprintf(outfile, "# Margo diagnostics\n");
    fprintf(outfile, "# %s\n", ctime(&ltime));
    fprintf(outfile, "# <stat>\t<avg>\t<min>\t<max>\t<count>\n");
1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170
    print_diag_data(outfile, "trigger_elapsed", 
        "Time consumed by HG_Trigger()", 
        &mid->diag_trigger_elapsed);
    print_diag_data(outfile, "progress_elapsed_zero_timeout", 
        "Time consumed by HG_Progress() when called with timeout==0", 
        &mid->diag_progress_elapsed_zero_timeout);
    print_diag_data(outfile, "progress_elapsed_nonzero_timeout", 
        "Time consumed by HG_Progress() when called with timeout!=0", 
        &mid->diag_progress_elapsed_nonzero_timeout);
    print_diag_data(outfile, "progress_timeout_value", 
        "Timeout values passed to HG_Progress()", 
        &mid->diag_progress_timeout_value);
Philip Carns's avatar
Philip Carns committed
1171 1172 1173 1174 1175 1176

    if(outfile != stdout)
        fclose(outfile);
    
    return;
}