diff options
Diffstat (limited to 'pkev.h')
| -rw-r--r-- | pkev.h | 76 |
1 files changed, 50 insertions, 26 deletions
@@ -1,6 +1,8 @@ #ifndef PK_EV_H #define PK_EV_H +#include "./pkmem.h" /* deleteme */ + #include <stdint.h> typedef uint64_t pk_ev_mgr_id_T; @@ -10,7 +12,7 @@ typedef uint64_t pk_ev_cb_id_T; // TODO re-think threading // note: pk_ev_init() is NOT thread-safe -void pk_ev_init(); +void pk_ev_init(struct pk_membucket *bkt); // note: pk_ev_teardown() is NOT thread-safe void pk_ev_teardown(); @@ -55,6 +57,14 @@ void pk_ev_unregister_cb(pk_ev_mgr_id_T evmgr, pk_ev_id_T evid, pk_ev_cb_id_T cb # define PK_EV_GROW_RATIO 1.5 #endif +#ifndef PK_EV_MEM_ALLOC +# define PK_EV_MEM_ALLOC(sz, alignment, bkt) pk_new(sz, alignment, bkt) +#endif + +#ifndef PK_EV_MEM_FREE +# define PK_EV_MEM_FREE(ptr, sz, bkt) pk_delete(ptr, sz, bkt) +#endif + struct pk_ev_cb { pk_ev_cb_fn *cb; void *user_cb_data; @@ -81,16 +91,19 @@ struct pk_ev_mstr { atomic_uint_fast64_t rn_mgrs; struct pk_ev_mgr **mgrs; mtx_t *mtxs; + struct pk_membucket *bkt; }; struct pk_ev_mstr pk_ev_mstr; void -pk_ev_init() +pk_ev_init(struct pk_membucket* bkt) { int i; - pk_ev_mstr.mgrs = (struct pk_ev_mgr **)malloc(sizeof(void *) * PK_EV_INIT_MGR_COUNT); - pk_ev_mstr.mtxs = (mtx_t*)malloc(sizeof(mtx_t) * PK_EV_INIT_MGR_COUNT); + pk_ev_mstr.bkt = bkt; + pk_ev_mstr.mgrs = (struct pk_ev_mgr **)PK_EV_MEM_ALLOC(sizeof(void *) * + PK_EV_INIT_MGR_COUNT, alignof(void *), bkt); + pk_ev_mstr.mtxs = (mtx_t*)PK_EV_MEM_ALLOC(sizeof(mtx_t) * PK_EV_INIT_MGR_COUNT, alignof(mtx_t), bkt); memset(pk_ev_mstr.mgrs, 0, sizeof(void *) * PK_EV_INIT_MGR_COUNT); memset(pk_ev_mstr.mtxs, 0, sizeof(mtx_t) * PK_EV_INIT_MGR_COUNT); for (i = 0; i < PK_EV_INIT_MGR_COUNT; ++i) { @@ -100,37 +113,44 @@ pk_ev_init() atomic_store(&pk_ev_mstr.rn_mgrs, PK_EV_INIT_MGR_COUNT); } +size_t +pk_ev_inner_calc_sz(uint64_t ev_count, uint64_t cb_count, size_t *sz_ev_list, size_t *sz_ev_cb_list) +{ + // base sizes + size_t l_sz_ev_list = sizeof(struct pk_ev) * ev_count; + size_t l_sz_ev_cb_list = sizeof(struct pk_ev_cb) * cb_count; + if (sz_ev_list != nullptr) *sz_ev_list = l_sz_ev_list; + if (sz_ev_cb_list != nullptr) *sz_ev_cb_list = l_sz_ev_cb_list; + + size_t ret = sizeof(struct pk_ev_mgr); + ret += l_sz_ev_list; + ret += l_sz_ev_cb_list * ev_count; + return ret; +} + void pk_ev_teardown() { long unsigned int i; - for (i = 0; i < pk_ev_mstr.rn_mgrs; ++i) { - if ((atomic_load(&pk_ev_mstr.rn_mgrs) & (1lu << i)) == 0) continue; + for (i = 0; i < atomic_load(&pk_ev_mstr.rn_mgrs); ++i) { + if ((atomic_load(&pk_ev_mstr.flg_mgrs) & (1lu << i)) == 0) continue; mtx_lock(&pk_ev_mstr.mtxs[i]); - free(pk_ev_mstr.mgrs[i]); + size_t sz = pk_ev_inner_calc_sz( + atomic_load(&pk_ev_mstr.mgrs[i]->rn_ev), + atomic_load(&pk_ev_mstr.mgrs[i]->rn_cb), + NULL, NULL + ); + PK_EV_MEM_FREE(pk_ev_mstr.mgrs[i], sz, pk_ev_mstr.bkt); pk_ev_mstr.mgrs[i] = NULL; mtx_unlock(&pk_ev_mstr.mtxs[i]); mtx_destroy(&pk_ev_mstr.mtxs[i]); } - free(pk_ev_mstr.mgrs); - free(pk_ev_mstr.mtxs); + PK_EV_MEM_FREE(pk_ev_mstr.mgrs, sizeof(void *) * atomic_load(&pk_ev_mstr.rn_mgrs), pk_ev_mstr.bkt); + PK_EV_MEM_FREE(pk_ev_mstr.mtxs, sizeof(mtx_t) * atomic_load(&pk_ev_mstr.rn_mgrs), pk_ev_mstr.bkt); pk_ev_mstr.mgrs = NULL; pk_ev_mstr.mtxs = NULL; } -size_t -pk_ev_inner_calc_sz(uint64_t ev_count, uint64_t cb_count, size_t *sz_ev_list, size_t *sz_ev_cb_list) -{ - // base sizes - if (sz_ev_list != nullptr) *sz_ev_list = sizeof(struct pk_ev) * ev_count; - if (sz_ev_cb_list != nullptr) *sz_ev_cb_list = sizeof(struct pk_ev_cb) * cb_count; - - size_t ret = sizeof(struct pk_ev_mgr); - if (sz_ev_list != nullptr) ret += *sz_ev_list; - if (sz_ev_cb_list != nullptr) ret += *sz_ev_cb_list * ev_count; - return ret; -} - static struct pk_ev_mgr* pk_ev_inner_ev_mgr_create(uint64_t ev_count, uint64_t cb_count) { @@ -143,7 +163,7 @@ pk_ev_inner_ev_mgr_create(uint64_t ev_count, uint64_t cb_count) size_t sz = pk_ev_inner_calc_sz(ev_count, cb_count, &sz_ev_list, &sz_ev_cb_list); size_t sz_offset; - struct pk_ev_mgr *mgr = (struct pk_ev_mgr*)malloc(sz); + struct pk_ev_mgr *mgr = (struct pk_ev_mgr*)PK_EV_MEM_ALLOC(sz, alignof(struct pk_ev_mgr), pk_ev_mstr.bkt); if (mgr == NULL) goto early_exit; mgr->ev = (struct pk_ev*)(((char *)mgr) + sizeof(struct pk_ev_mgr)); @@ -219,8 +239,10 @@ recreate: void pk_ev_destroy_mgr(pk_ev_mgr_id_T evmgr) { + assert(evmgr < pk_ev_mstr.rn_mgrs); mtx_lock(&pk_ev_mstr.mtxs[evmgr]); - free(pk_ev_mstr.mgrs[evmgr]); + size_t old_sz = pk_ev_inner_calc_sz(pk_ev_mstr.mgrs[evmgr]->rn_ev, pk_ev_mstr.mgrs[evmgr]->rn_cb, NULL, NULL); + PK_EV_MEM_FREE(pk_ev_mstr.mgrs[evmgr], old_sz, pk_ev_mstr.bkt); pk_ev_mstr.mgrs[evmgr] = NULL; mtx_unlock(&pk_ev_mstr.mtxs[evmgr]); } @@ -242,7 +264,8 @@ pk_ev_register_ev(pk_ev_mgr_id_T evmgr, void *user_ev_data) } mgr = pk_ev_inner_ev_mgr_create(new_size, pk_ev_mstr.mgrs[evmgr]->rn_cb); pk_ev_inner_ev_mgr_clone(pk_ev_mstr.mgrs[evmgr], mgr); - free(pk_ev_mstr.mgrs[evmgr]); + size_t old_sz = pk_ev_inner_calc_sz(pk_ev_mstr.mgrs[evmgr]->rn_ev, pk_ev_mstr.mgrs[evmgr]->rn_cb, NULL, NULL); + PK_EV_MEM_FREE(pk_ev_mstr.mgrs[evmgr], old_sz, pk_ev_mstr.bkt); pk_ev_mstr.mgrs[evmgr] = mgr; } id = pk_ev_mstr.mgrs[evmgr]->n_ev++; @@ -278,6 +301,7 @@ pk_ev_register_cb(pk_ev_mgr_id_T evmgr, pk_ev_id_T evid, pk_ev_cb_fn *cb, void * if (found == false) { mtx_lock(&pk_ev_mstr.mtxs[evmgr]); if (pk_ev_mstr.mgrs[evmgr]->ev[evid].right_ev_cbs == pk_ev_mstr.mgrs[evmgr]->rn_cb) { + size_t old_sz = pk_ev_inner_calc_sz(pk_ev_mstr.mgrs[evmgr]->rn_ev, pk_ev_mstr.mgrs[evmgr]->rn_cb, NULL, NULL); new_size = PK_MAX(2, PK_MIN(255, pk_ev_mstr.mgrs[evmgr]->rn_cb * PK_EV_GROW_RATIO)); if (new_size == pk_ev_mstr.mgrs[evmgr]->rn_cb) { PK_LOG_ERR("[pkev.h] need more room, but failed to grow cb count.\n"); @@ -286,7 +310,7 @@ pk_ev_register_cb(pk_ev_mgr_id_T evmgr, pk_ev_id_T evid, pk_ev_cb_fn *cb, void * } mgr = pk_ev_inner_ev_mgr_create(pk_ev_mstr.mgrs[evmgr]->rn_ev, new_size); pk_ev_inner_ev_mgr_clone(pk_ev_mstr.mgrs[evmgr], mgr); - free(pk_ev_mstr.mgrs[evmgr]); + PK_EV_MEM_FREE(pk_ev_mstr.mgrs[evmgr], old_sz, pk_ev_mstr.bkt); pk_ev_mstr.mgrs[evmgr] = mgr; mgr = nullptr; } |
