margo.c 14 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11

/*
 * (C) 2015 The University of Chicago
 * 
 * See COPYRIGHT in top-level directory.
 */

#include <assert.h>
#include <unistd.h>
#include <errno.h>
#include <abt.h>
12
#include <abt-snoozer.h>
13
#include <time.h>
Philip Carns's avatar
Philip Carns committed
14
#include <math.h>
15 16

#include "margo.h"
17
#include "margo-timer.h"
Philip Carns's avatar
Philip Carns committed
18
#include "utlist.h"
19

20 21 22 23 24
/* TODO: including core.h for cancel definition, presumably this will be 
 * available in top level later?
 */
#include <mercury_core.h>

25 26
#define MERCURY_PROGRESS_TIMEOUT_UB 100 /* 100 milliseconds */

27 28
struct margo_instance
{
29
    /* provided by caller */
30 31
    hg_context_t *hg_context;
    hg_class_t *hg_class;
32 33 34
    ABT_pool handler_pool;
    ABT_pool progress_pool;

35
    /* internal to margo for this particular instance */
36 37
    ABT_thread hg_progress_tid;
    int hg_progress_shutdown_flag;
38 39 40 41 42 43 44

    /* control logic for callers waiting on margo to be finalized */
    int finalize_flag;
    int finalize_waiters_in_progress_pool;
    ABT_mutex finalize_mutex;
    ABT_cond finalize_cond;

45 46 47 48 49 50 51 52 53 54 55 56
    int table_index;
};

struct margo_handler_mapping
{
    hg_class_t *class;
    margo_instance_id mid;
};

#define MAX_HANDLER_MAPPING 8
static int handler_mapping_table_size = 0;
static struct margo_handler_mapping handler_mapping_table[MAX_HANDLER_MAPPING] = {0};
57

58
static void hg_progress_fn(void* foo);
59
static int margo_xstream_is_in_progress_pool(margo_instance_id mid);
60 61 62 63 64 65 66 67

struct handler_entry
{
    void* fn;
    hg_handle_t handle;
    struct handler_entry *next; 
};

68 69
margo_instance_id margo_init(ABT_pool progress_pool, ABT_pool handler_pool,
    hg_context_t *hg_context, hg_class_t *hg_class)
70 71
{
    int ret;
72 73 74
    struct margo_instance *mid;

    if(handler_mapping_table_size >= MAX_HANDLER_MAPPING)
75
        return(MARGO_INSTANCE_NULL);
76 77 78

    mid = malloc(sizeof(*mid));
    if(!mid)
79
        return(MARGO_INSTANCE_NULL);
80
    memset(mid, 0, sizeof(*mid));
81

82 83 84
    ABT_mutex_create(&mid->finalize_mutex);
    ABT_cond_create(&mid->finalize_cond);

85 86
    mid->progress_pool = progress_pool;
    mid->handler_pool = handler_pool;
87 88
    mid->hg_class = hg_class;
    mid->hg_context = hg_context;
89

90 91 92 93 94 95 96 97
    ret = margo_timer_instance_init(mid);
    if(ret != 0)
    {
        fprintf(stderr, "Error: margo_timer_instance_init()\n");
        free(mid);
        return(MARGO_INSTANCE_NULL);
    }

98
    ret = ABT_thread_create(mid->progress_pool, hg_progress_fn, mid, 
99
        ABT_THREAD_ATTR_NULL, &mid->hg_progress_tid);
100 101 102
    if(ret != 0)
    {
        fprintf(stderr, "Error: ABT_thread_create()\n");
103
        free(mid);
104
        return(MARGO_INSTANCE_NULL);
105 106
    }

107 108 109 110 111 112
    handler_mapping_table[handler_mapping_table_size].mid = mid;
    handler_mapping_table[handler_mapping_table_size].class = mid->hg_class;
    mid->table_index = handler_mapping_table_size;
    handler_mapping_table_size++;

    return mid;
113 114
}

115
void margo_finalize(margo_instance_id mid)
116
{
117 118
    int i;

119
    /* tell progress thread to wrap things up */
120
    mid->hg_progress_shutdown_flag = 1;
121 122

    /* wait for it to shutdown cleanly */
123 124
    ABT_thread_join(mid->hg_progress_tid);
    ABT_thread_free(&mid->hg_progress_tid);
125

126 127 128 129 130
    for(i=mid->table_index; i<(handler_mapping_table_size-1); i++)
    {
        handler_mapping_table[i] = handler_mapping_table[i+1];
    }
    handler_mapping_table_size--;
131

132 133 134 135 136 137 138 139 140 141 142 143 144
    ABT_mutex_lock(mid->finalize_mutex);
    mid->finalize_flag = 1;
    ABT_cond_broadcast(mid->finalize_cond);
    ABT_mutex_unlock(mid->finalize_mutex);

    /* TODO: yuck, there is a race here if someone was really waiting for
     * finalize; we can't destroy the data structures out from under them.
     * We could fix this by reference counting so that the last caller
     * (whether a finalize() caller or wait_for_finalize() caller) knows it
     * is safe to turn off the lights on their way out.  For now we just leak 
     * a small amount of memory.
     */
#if 0
145
    margo_timer_instance_finalize(mid);
146

147 148 149 150 151 152 153 154 155 156 157 158 159 160 161
    ABT_mutex_free(&mid->finalize_mutex);
    ABT_cond_free(&mid->finalize_cond);
    free(mid);
#endif

    return;
}

void margo_wait_for_finalize(margo_instance_id mid)
{
    int in_pool = 0;

    /* Is this waiter in the same pool as the pool running the progress
     * thread?
     */
162
    if(margo_xstream_is_in_progress_pool(mid))
163 164 165 166 167 168 169 170 171 172 173
        in_pool = 1;

    ABT_mutex_lock(mid->finalize_mutex);

        mid->finalize_waiters_in_progress_pool += in_pool;
            
        while(!mid->finalize_flag)
            ABT_cond_wait(mid->finalize_cond, mid->finalize_mutex);

    ABT_mutex_unlock(mid->finalize_mutex);
    
174 175 176 177
    return;
}

/* dedicated thread function to drive Mercury progress */
178
static void hg_progress_fn(void* foo)
179 180 181
{
    int ret;
    unsigned int actual_count;
182
    struct margo_instance *mid = (struct margo_instance *)foo;
183
    size_t size;
184 185
    unsigned int hg_progress_timeout = MERCURY_PROGRESS_TIMEOUT_UB;
    double next_timer_exp;
186

187
    while(!mid->hg_progress_shutdown_flag)
188 189
    {
        do {
190
            ret = HG_Trigger(mid->hg_context, 0, 1, &actual_count);
191
        } while((ret == HG_SUCCESS) && actual_count && !mid->hg_progress_shutdown_flag);
192

193
        if(!mid->hg_progress_shutdown_flag)
194
        {
195 196
            ABT_mutex_lock(mid->finalize_mutex);

197
            ABT_pool_get_total_size(mid->progress_pool, &size);
198 199 200 201 202 203
            /* Are there any other threads executing in this pool that are *not*
             * blocked on margo_wait_for_finalize()?  If so then, we can't
             * sleep here or else those threads will not get a chance to
             * execute.
             */
            if(size > mid->finalize_waiters_in_progress_pool)
204
            {
205
                ABT_mutex_unlock(mid->finalize_mutex);
206
                HG_Progress(mid->hg_context, 0);
207 208 209 210
                ABT_thread_yield();
            }
            else
            {
211
                ABT_mutex_unlock(mid->finalize_mutex);
212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230

                ret = margo_timer_get_next_expiration(mid, &next_timer_exp);
                if(ret == 0)
                {
                    /* there is a queued timer, don't block long enough
                     * to keep this timer waiting
                     */
                    if(next_timer_exp >= 0.0)
                    {
                        next_timer_exp *= 1000; /* convert to milliseconds */
                        if(next_timer_exp < MERCURY_PROGRESS_TIMEOUT_UB)
                            hg_progress_timeout = (unsigned int)next_timer_exp;
                    }
                    else
                    {
                        hg_progress_timeout = 0;
                    }
                }
                HG_Progress(mid->hg_context, hg_progress_timeout);
231 232
            }
        }
233

234
        /* check for any expired timers */
235
        margo_check_timers(mid);
236 237
    }

238
    return;
239 240
}

241
ABT_pool* margo_get_handler_pool(margo_instance_id mid)
242
{
243
    return(&mid->handler_pool);
244 245
}

246 247 248 249 250 251 252 253 254 255 256
hg_context_t* margo_get_context(margo_instance_id mid)
{
    return(mid->hg_context);
}

hg_class_t* margo_get_class(margo_instance_id mid)
{
    return(mid->hg_class);
}


Jonathan Jenkins's avatar
Jonathan Jenkins committed
257
static hg_return_t margo_cb(const struct hg_cb_info *info)
258 259 260 261 262 263 264 265 266 267
{
    hg_return_t hret = info->ret;

    ABT_eventual *eventual = info->arg;
    /* propagate return code out through eventual */
    ABT_eventual_set(*eventual, &hret, sizeof(hret));
    
    return(HG_SUCCESS);
}

268 269 270
typedef struct
{
    hg_handle_t handle;
Shane Snyder's avatar
Shane Snyder committed
271
} margo_forward_timeout_cb_dat;
272 273 274

static void margo_forward_timeout_cb(void *arg)
{
Shane Snyder's avatar
Shane Snyder committed
275 276
    margo_forward_timeout_cb_dat *timeout_cb_dat =
        (margo_forward_timeout_cb_dat *)arg;
277 278

    /* cancel the Mercury op if the forward timed out */
Shane Snyder's avatar
Shane Snyder committed
279
    HG_Core_cancel(timeout_cb_dat->handle);
280 281 282
    return;
}

283 284 285 286 287 288
hg_return_t margo_forward_timed(
    margo_instance_id mid,
    hg_handle_t handle,
    void *in_struct,
    double timeout_ms)
{
Shane Snyder's avatar
Shane Snyder committed
289
    int ret;
290 291 292
    hg_return_t hret = HG_TIMEOUT;
    ABT_eventual eventual;
    hg_return_t* waited_hret;
Shane Snyder's avatar
Shane Snyder committed
293 294
    margo_timer_t forward_timer;
    margo_forward_timeout_cb_dat timeout_cb_dat;
295 296 297 298 299 300 301

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

Shane Snyder's avatar
Shane Snyder committed
302 303
    /* set a timer object to expire when this forward times out */
    timeout_cb_dat.handle = handle;
304
    margo_timer_init(mid, &forward_timer, margo_forward_timeout_cb,
Shane Snyder's avatar
Shane Snyder committed
305
        &timeout_cb_dat, timeout_ms);
306 307 308 309 310 311 312 313

    hret = HG_Forward(handle, margo_cb, &eventual, in_struct);
    if(hret == 0)
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

314 315
    /* remove timer if it is still in place (i.e., not timed out) */
    if(hret != HG_TIMEOUT)
316
        margo_timer_destroy(mid, &forward_timer);
317 318 319 320

    ABT_eventual_free(&eventual);

    return(hret);
321 322 323 324

}


325
hg_return_t margo_forward(
326
    margo_instance_id mid,
327 328 329 330 331 332 333 334 335 336 337 338 339 340
    hg_handle_t handle,
    void *in_struct)
{
    hg_return_t hret = HG_TIMEOUT;
    ABT_eventual eventual;
    int ret;
    hg_return_t* waited_hret;

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

Jonathan Jenkins's avatar
Jonathan Jenkins committed
341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369
    hret = HG_Forward(handle, margo_cb, &eventual, in_struct);
    if(hret == 0)
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

    ABT_eventual_free(&eventual);

    return(hret);
}

hg_return_t margo_respond(
    margo_instance_id mid,
    hg_handle_t handle,
    void *out_struct)
{
    hg_return_t hret = HG_TIMEOUT;
    ABT_eventual eventual;
    int ret;
    hg_return_t* waited_hret;

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);
    }

    hret = HG_Respond(handle, margo_cb, &eventual, out_struct);
370 371 372 373 374 375 376 377 378 379 380
    if(hret == 0)
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

    ABT_eventual_free(&eventual);

    return(hret);
}

381

382
static hg_return_t margo_bulk_transfer_cb(const struct hg_cb_info *info)
383
{
384 385
    hg_return_t hret = info->ret;
    ABT_eventual *eventual = info->arg;
386 387 388 389 390 391 392

    /* propagate return code out through eventual */
    ABT_eventual_set(*eventual, &hret, sizeof(hret));
    
    return(HG_SUCCESS);
}

Philip Carns's avatar
Philip Carns committed
393 394
struct lookup_cb_evt
{
395 396
    hg_return_t nret;
    hg_addr_t addr;
Philip Carns's avatar
Philip Carns committed
397 398
};

399
static hg_return_t margo_addr_lookup_cb(const struct hg_cb_info *info)
Philip Carns's avatar
Philip Carns committed
400 401
{
    struct lookup_cb_evt evt;
402 403
    evt.nret = info->ret;
    evt.addr = info->info.lookup.addr;
Philip Carns's avatar
Philip Carns committed
404

405
    ABT_eventual *eventual = info->arg;
Philip Carns's avatar
Philip Carns committed
406 407 408 409

    /* propagate return code out through eventual */
    ABT_eventual_set(*eventual, &evt, sizeof(evt));
    
410
    return(HG_SUCCESS);
Philip Carns's avatar
Philip Carns committed
411 412 413
}


414
hg_return_t margo_addr_lookup(
415
    margo_instance_id mid,
416
    hg_context_t *context,
Philip Carns's avatar
Philip Carns committed
417
    const char   *name,
418
    hg_addr_t    *addr)
419
{
420
    hg_return_t nret;
Philip Carns's avatar
Philip Carns committed
421
    struct lookup_cb_evt *evt;
422 423 424
    ABT_eventual eventual;
    int ret;

Philip Carns's avatar
Philip Carns committed
425
    ret = ABT_eventual_create(sizeof(*evt), &eventual);
426 427 428 429 430
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

431 432
    nret = HG_Addr_lookup(context, margo_addr_lookup_cb,
        &eventual, name, HG_OP_ID_IGNORE);
433 434
    if(nret == 0)
    {
Philip Carns's avatar
Philip Carns committed
435 436 437
        ABT_eventual_wait(eventual, (void**)&evt);
        *addr = evt->addr;
        nret = evt->nret;
438 439 440 441 442 443 444
    }

    ABT_eventual_free(&eventual);

    return(nret);
}

445
hg_return_t margo_bulk_transfer(
446
    margo_instance_id mid,
447
    hg_context_t *context,
448
    hg_bulk_op_t op,
449
    hg_addr_t origin_addr,
450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480
    hg_bulk_t origin_handle,
    size_t origin_offset,
    hg_bulk_t local_handle,
    size_t local_offset,
    size_t size)
{
    hg_return_t hret = HG_TIMEOUT;
    hg_return_t *waited_hret;
    ABT_eventual eventual;
    int ret;

    ret = ABT_eventual_create(sizeof(hret), &eventual);
    if(ret != 0)
    {
        return(HG_NOMEM_ERROR);        
    }

    hret = HG_Bulk_transfer(context, margo_bulk_transfer_cb, &eventual, op, 
        origin_addr, origin_handle, origin_offset, local_handle, local_offset,
        size, HG_OP_ID_IGNORE);
    if(hret == 0)
    {
        ABT_eventual_wait(eventual, (void**)&waited_hret);
        hret = *waited_hret;
    }

    ABT_eventual_free(&eventual);

    return(hret);
}

481 482
typedef struct
{
483
    margo_instance_id mid;
484 485
    ABT_mutex mutex;
    ABT_cond cond;
486
    int is_asleep;
487 488 489 490 491 492 493
} margo_thread_sleep_cb_dat;

static void margo_thread_sleep_cb(void *arg)
{
    margo_thread_sleep_cb_dat *sleep_cb_dat =
        (margo_thread_sleep_cb_dat *)arg;

494 495 496 497 498
    /* decrement number of waiting threads */
    ABT_mutex_lock(sleep_cb_dat->mid->finalize_mutex);
    sleep_cb_dat->mid->finalize_waiters_in_progress_pool--;
    ABT_mutex_unlock(sleep_cb_dat->mid->finalize_mutex);

499 500
    /* wake up the sleeping thread */
    ABT_mutex_lock(sleep_cb_dat->mutex);
501
    sleep_cb_dat->is_asleep = 0;
502 503 504 505 506 507 508
    ABT_cond_signal(sleep_cb_dat->cond);
    ABT_mutex_unlock(sleep_cb_dat->mutex);

    return;
}

void margo_thread_sleep(
509
    margo_instance_id mid,
510 511
    double timeout_ms)
{
512
    int in_pool = 0;
513 514 515 516
    margo_timer_t sleep_timer;
    margo_thread_sleep_cb_dat sleep_cb_dat;

    /* set data needed for sleep callback */
517
    sleep_cb_dat.mid = mid;
518 519
    ABT_mutex_create(&(sleep_cb_dat.mutex));
    ABT_cond_create(&(sleep_cb_dat.cond));
520
    sleep_cb_dat.is_asleep = 1;
521 522

    /* initialize the sleep timer */
523
    margo_timer_init(mid, &sleep_timer, margo_thread_sleep_cb,
524 525
        &sleep_cb_dat, timeout_ms);

526 527 528
    if(margo_xstream_is_in_progress_pool(mid))
        in_pool = 1;

529 530
    /* increment number of waiting threads */
    ABT_mutex_lock(mid->finalize_mutex);
531
    mid->finalize_waiters_in_progress_pool += in_pool;
532 533
    ABT_mutex_unlock(mid->finalize_mutex);

534 535
    /* yield thread for specified timeout */
    ABT_mutex_lock(sleep_cb_dat.mutex);
536 537
    while(sleep_cb_dat.is_asleep)
        ABT_cond_wait(sleep_cb_dat.cond, sleep_cb_dat.mutex);
538 539 540 541 542
    ABT_mutex_unlock(sleep_cb_dat.mutex);

    return;
}

543
margo_instance_id margo_hg_class_to_instance(hg_class_t *cl)
544 545 546 547 548
{
    int i;

    for(i=0; i<handler_mapping_table_size; i++)
    {
549
        if(handler_mapping_table[i].class == cl)
550 551 552 553
            return(handler_mapping_table[i].mid);
    }
    return(NULL);
}
554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571

/* returns 1 if current xstream is in the progress pool, 0 if not */
static int margo_xstream_is_in_progress_pool(margo_instance_id mid)
{
    int ret;
    ABT_xstream xstream;
    ABT_pool pool;

    ret = ABT_xstream_self(&xstream);
    assert(ret == ABT_SUCCESS);
    ret = ABT_xstream_get_main_pools(xstream, 1, &pool);
    assert(ret == ABT_SUCCESS);

    if(pool == mid->progress_pool)
        return(1);
    else
        return(0);
}