version 1.10.2.11, 2012/08/02 13:45:02
|
version 1.13.2.1, 2012/08/22 10:33:45
|
Line 46 SUCH DAMAGE.
|
Line 46 SUCH DAMAGE.
|
#include "global.h" |
#include "global.h" |
|
|
|
|
#pragma GCC visibility push(hidden) | /* |
| * sched_useTask() - Get and init new task |
| * |
| * @root = root task |
| * return: NULL error or !=NULL prepared task |
| */ |
inline sched_task_t * |
inline sched_task_t * |
_sched_useTask(sched_root_task_t * __restrict root) | sched_useTask(sched_root_task_t * __restrict root) |
{ |
{ |
sched_task_t *task, *tmp; |
sched_task_t *task, *tmp; |
|
|
Line 79 _sched_useTask(sched_root_task_t * __restrict root)
|
Line 83 _sched_useTask(sched_root_task_t * __restrict root)
|
return task; |
return task; |
} |
} |
|
|
|
/* |
|
* sched_unuseTask() - Unlock and put task to unuse queue |
|
* |
|
* @task = task |
|
* return: always is NULL |
|
*/ |
inline sched_task_t * |
inline sched_task_t * |
_sched_unuseTask(sched_task_t * __restrict task) | sched_unuseTask(sched_task_t * __restrict task) |
{ |
{ |
TASK_UNLOCK(task); |
TASK_UNLOCK(task); |
TASK_TYPE(task) = taskUNUSE; |
TASK_TYPE(task) = taskUNUSE; |
Line 96 _sched_unuseTask(sched_task_t * __restrict task)
|
Line 106 _sched_unuseTask(sched_task_t * __restrict task)
|
return task; |
return task; |
} |
} |
|
|
#pragma GCC visibility pop | void * |
| _sched_threadJoin(sched_task_t *task) |
| { |
| void *ret = NULL; |
|
|
|
if (!task) |
|
return NULL; |
|
|
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_join((pthread_t) TASK_VAL(task), &ret); |
|
TASK_ROOT(task)->root_ret = ret; |
|
#endif |
|
|
|
return NULL; |
|
} |
|
|
/* |
/* |
|
* sched_taskExit() - Exit routine for scheduler task, explicit required for thread tasks |
|
* |
|
* @task = current task |
|
* @retcode = return code |
|
* return: return code |
|
*/ |
|
inline void * |
|
sched_taskExit(sched_task_t *task, intptr_t retcode) |
|
{ |
|
if (!task || !TASK_ROOT(task)) |
|
return (void*) -1; |
|
|
|
if (TASK_ROOT(task)->root_hooks.hook_exec.exit) |
|
TASK_ROOT(task)->root_hooks.hook_exec.exit(task, (void*) retcode); |
|
|
|
TASK_ROOT(task)->root_ret = (void*) retcode; |
|
|
|
#ifdef HAVE_LIBPTHREAD |
|
if (TASK_TYPE(task) == taskTHREAD) { |
|
if (TASK_FLAG(task) == PTHREAD_CREATE_JOINABLE) /* joinable thread */ |
|
schedTask(TASK_ROOT(task), _sched_threadJoin, TASK_ARG(task), |
|
TASK_VAL(task), TASK_DATA(task), TASK_DATLEN(task)); |
|
sched_unuseTask(task); |
|
pthread_exit((void*) retcode); |
|
} |
|
#endif |
|
|
|
return (void*) retcode; |
|
} |
|
|
|
|
|
/* |
* schedRead() - Add READ I/O task to scheduler queue |
* schedRead() - Add READ I/O task to scheduler queue |
* |
* |
* @root = root task |
* @root = root task |
Line 121 schedRead(sched_root_task_t * __restrict root, sched_t
|
Line 176 schedRead(sched_root_task_t * __restrict root, sched_t
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 148 schedRead(sched_root_task_t * __restrict root, sched_t
|
Line 203 schedRead(sched_root_task_t * __restrict root, sched_t
|
pthread_mutex_unlock(&root->root_mtx[taskREAD]); |
pthread_mutex_unlock(&root->root_mtx[taskREAD]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 175 schedWrite(sched_root_task_t * __restrict root, sched_
|
Line 230 schedWrite(sched_root_task_t * __restrict root, sched_
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 202 schedWrite(sched_root_task_t * __restrict root, sched_
|
Line 257 schedWrite(sched_root_task_t * __restrict root, sched_
|
pthread_mutex_unlock(&root->root_mtx[taskWRITE]); |
pthread_mutex_unlock(&root->root_mtx[taskWRITE]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 229 schedNode(sched_root_task_t * __restrict root, sched_t
|
Line 284 schedNode(sched_root_task_t * __restrict root, sched_t
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 256 schedNode(sched_root_task_t * __restrict root, sched_t
|
Line 311 schedNode(sched_root_task_t * __restrict root, sched_t
|
pthread_mutex_unlock(&root->root_mtx[taskNODE]); |
pthread_mutex_unlock(&root->root_mtx[taskNODE]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 283 schedProc(sched_root_task_t * __restrict root, sched_t
|
Line 338 schedProc(sched_root_task_t * __restrict root, sched_t
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 310 schedProc(sched_root_task_t * __restrict root, sched_t
|
Line 365 schedProc(sched_root_task_t * __restrict root, sched_t
|
pthread_mutex_unlock(&root->root_mtx[taskPROC]); |
pthread_mutex_unlock(&root->root_mtx[taskPROC]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 341 schedUser(sched_root_task_t * __restrict root, sched_t
|
Line 396 schedUser(sched_root_task_t * __restrict root, sched_t
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 368 schedUser(sched_root_task_t * __restrict root, sched_t
|
Line 423 schedUser(sched_root_task_t * __restrict root, sched_t
|
pthread_mutex_unlock(&root->root_mtx[taskUSER]); |
pthread_mutex_unlock(&root->root_mtx[taskUSER]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
#endif |
#endif |
Line 396 schedSignal(sched_root_task_t * __restrict root, sched
|
Line 451 schedSignal(sched_root_task_t * __restrict root, sched
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 423 schedSignal(sched_root_task_t * __restrict root, sched
|
Line 478 schedSignal(sched_root_task_t * __restrict root, sched
|
pthread_mutex_unlock(&root->root_mtx[taskSIGNAL]); |
pthread_mutex_unlock(&root->root_mtx[taskSIGNAL]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 450 schedAlarm(sched_root_task_t * __restrict root, sched_
|
Line 505 schedAlarm(sched_root_task_t * __restrict root, sched_
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 477 schedAlarm(sched_root_task_t * __restrict root, sched_
|
Line 532 schedAlarm(sched_root_task_t * __restrict root, sched_
|
pthread_mutex_unlock(&root->root_mtx[taskALARM]); |
pthread_mutex_unlock(&root->root_mtx[taskALARM]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 505 schedAIO(sched_root_task_t * __restrict root, sched_ta
|
Line 560 schedAIO(sched_root_task_t * __restrict root, sched_ta
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 532 schedAIO(sched_root_task_t * __restrict root, sched_ta
|
Line 587 schedAIO(sched_root_task_t * __restrict root, sched_ta
|
pthread_mutex_unlock(&root->root_mtx[taskAIO]); |
pthread_mutex_unlock(&root->root_mtx[taskAIO]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 668 schedLIO(sched_root_task_t * __restrict root, sched_ta
|
Line 723 schedLIO(sched_root_task_t * __restrict root, sched_ta
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 695 schedLIO(sched_root_task_t * __restrict root, sched_ta
|
Line 750 schedLIO(sched_root_task_t * __restrict root, sched_ta
|
pthread_mutex_unlock(&root->root_mtx[taskLIO]); |
pthread_mutex_unlock(&root->root_mtx[taskLIO]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 745 schedLIORead(sched_root_task_t * __restrict root, sche
|
Line 800 schedLIORead(sched_root_task_t * __restrict root, sche
|
for (i = 0; i < nbufs; i++) |
for (i = 0; i < nbufs; i++) |
if (acb[i]) |
if (acb[i]) |
free(acb[i]); |
free(acb[i]); |
free(acb); | free(acb); |
return NULL; |
return NULL; |
} else |
} else |
memset(acb[i], 0, sizeof(struct aiocb)); |
memset(acb[i], 0, sizeof(struct aiocb)); |
Line 762 schedLIORead(sched_root_task_t * __restrict root, sche
|
Line 817 schedLIORead(sched_root_task_t * __restrict root, sche
|
|
|
if (lio_listio(LIO_NOWAIT, acb, nbufs, &sig)) { |
if (lio_listio(LIO_NOWAIT, acb, nbufs, &sig)) { |
LOGERR; |
LOGERR; |
|
for (i = 0; i < nbufs; i++) |
|
if (acb[i]) |
|
free(acb[i]); |
|
free(acb); |
return NULL; |
return NULL; |
} |
} |
|
|
Line 813 schedLIOWrite(sched_root_task_t * __restrict root, sch
|
Line 872 schedLIOWrite(sched_root_task_t * __restrict root, sch
|
for (i = 0; i < nbufs; i++) |
for (i = 0; i < nbufs; i++) |
if (acb[i]) |
if (acb[i]) |
free(acb[i]); |
free(acb[i]); |
free(acb); | free(acb); |
return NULL; |
return NULL; |
} else |
} else |
memset(acb[i], 0, sizeof(struct aiocb)); |
memset(acb[i], 0, sizeof(struct aiocb)); |
Line 830 schedLIOWrite(sched_root_task_t * __restrict root, sch
|
Line 889 schedLIOWrite(sched_root_task_t * __restrict root, sch
|
|
|
if (lio_listio(LIO_NOWAIT, acb, nbufs, &sig)) { |
if (lio_listio(LIO_NOWAIT, acb, nbufs, &sig)) { |
LOGERR; |
LOGERR; |
|
for (i = 0; i < nbufs; i++) |
|
if (acb[i]) |
|
free(acb[i]); |
|
free(acb); |
return NULL; |
return NULL; |
} |
} |
|
|
Line 861 schedTimer(sched_root_task_t * __restrict root, sched_
|
Line 924 schedTimer(sched_root_task_t * __restrict root, sched_
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 910 schedTimer(sched_root_task_t * __restrict root, sched_
|
Line 973 schedTimer(sched_root_task_t * __restrict root, sched_
|
pthread_mutex_unlock(&root->root_mtx[taskTIMER]); |
pthread_mutex_unlock(&root->root_mtx[taskTIMER]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 937 schedEvent(sched_root_task_t * __restrict root, sched_
|
Line 1000 schedEvent(sched_root_task_t * __restrict root, sched_
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 964 schedEvent(sched_root_task_t * __restrict root, sched_
|
Line 1027 schedEvent(sched_root_task_t * __restrict root, sched_
|
pthread_mutex_unlock(&root->root_mtx[taskEVENT]); |
pthread_mutex_unlock(&root->root_mtx[taskEVENT]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
|
|
|
|
/* |
/* |
* schedEventLo() - Add EVENT_Lo task to scheduler queue | * schedTask() - Add regular task to scheduler queue |
* |
* |
* @root = root task |
* @root = root task |
* @func = task execution function |
* @func = task execution function |
* @arg = 1st func argument |
* @arg = 1st func argument |
* @val = additional func argument | * @prio = regular task priority, 0 is hi priority for regular tasks |
* @opt_data = Optional data |
* @opt_data = Optional data |
* @opt_dlen = Optional data length |
* @opt_dlen = Optional data length |
* return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
*/ |
*/ |
sched_task_t * |
sched_task_t * |
schedEventLo(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, u_long val, | schedTask(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, u_long prio, |
void *opt_data, size_t opt_dlen) |
void *opt_data, size_t opt_dlen) |
{ |
{ |
sched_task_t *task; | sched_task_t *task, *tmp, *t = NULL; |
void *ptr; |
void *ptr; |
|
|
if (!root || !func) |
if (!root || !func) |
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
TASK_TYPE(task) = taskEVENT; | TASK_TYPE(task) = taskTASK; |
TASK_ROOT(task) = root; |
TASK_ROOT(task) = root; |
|
|
TASK_ARG(task) = arg; |
TASK_ARG(task) = arg; |
TASK_VAL(task) = val; | TASK_VAL(task) = prio; |
|
|
TASK_DATA(task) = opt_data; |
TASK_DATA(task) = opt_data; |
TASK_DATLEN(task) = opt_dlen; |
TASK_DATLEN(task) = opt_dlen; |
|
|
if (root->root_hooks.hook_add.eventlo) | if (root->root_hooks.hook_add.task) |
ptr = root->root_hooks.hook_add.eventlo(task, NULL); | ptr = root->root_hooks.hook_add.task(task, NULL); |
else |
else |
ptr = NULL; |
ptr = NULL; |
|
|
if (!ptr) { |
if (!ptr) { |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&root->root_mtx[taskEVENTLO]); | pthread_mutex_lock(&root->root_mtx[taskTASK]); |
#endif |
#endif |
TAILQ_INSERT_TAIL(&root->root_eventlo, TASK_ID(task), task_node); | TAILQ_FOREACH_SAFE(t, &root->root_task, task_node, tmp) |
| if (TASK_VAL(task) < TASK_VAL(t)) |
| break; |
| if (!t) |
| TAILQ_INSERT_TAIL(&root->root_task, TASK_ID(task), task_node); |
| else |
| TAILQ_INSERT_BEFORE(t, TASK_ID(task), task_node); |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_unlock(&root->root_mtx[taskEVENTLO]); | pthread_mutex_unlock(&root->root_mtx[taskTASK]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 1046 schedSuspend(sched_root_task_t * __restrict root, sche
|
Line 1115 schedSuspend(sched_root_task_t * __restrict root, sche
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 1073 schedSuspend(sched_root_task_t * __restrict root, sche
|
Line 1142 schedSuspend(sched_root_task_t * __restrict root, sche
|
pthread_mutex_unlock(&root->root_mtx[taskSUSPEND]); |
pthread_mutex_unlock(&root->root_mtx[taskSUSPEND]); |
#endif |
#endif |
} else |
} else |
task = _sched_unuseTask(task); | task = sched_unuseTask(task); |
|
|
return task; |
return task; |
} |
} |
Line 1100 schedCallOnce(sched_root_task_t * __restrict root, sch
|
Line 1169 schedCallOnce(sched_root_task_t * __restrict root, sch
|
return NULL; |
return NULL; |
|
|
/* get new task */ |
/* get new task */ |
if (!(task = _sched_useTask(root))) | if (!(task = sched_useTask(root))) |
return NULL; |
return NULL; |
|
|
task->task_func = func; |
task->task_func = func; |
Line 1115 schedCallOnce(sched_root_task_t * __restrict root, sch
|
Line 1184 schedCallOnce(sched_root_task_t * __restrict root, sch
|
|
|
ret = schedCall(task); |
ret = schedCall(task); |
|
|
_sched_unuseTask(task); | sched_unuseTask(task); |
return ret; |
return ret; |
} |
} |
|
|
|
/* |
|
* schedThread() - Add thread task to scheduler queue |
|
* |
|
* @root = root task |
|
* @func = task execution function |
|
* @arg = 1st func argument |
|
* @detach = Detach thread from scheduler, if !=0 |
|
* @opt_data = Optional data |
|
* @opt_dlen = Optional data length |
|
* return: NULL error or !=NULL new queued task |
|
*/ |
|
sched_task_t * |
|
schedThread(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, int detach, |
|
void *opt_data, size_t opt_dlen) |
|
{ |
|
#ifndef HAVE_LIBPTHREAD |
|
sched_SetErr(ENOTSUP, "Not supported thread tasks"); |
|
return NULL; |
|
#endif |
|
sched_task_t *task; |
|
void *ptr; |
|
pthread_attr_t attr; |
|
|
|
if (!root || !func) |
|
return NULL; |
|
|
|
/* get new task */ |
|
if (!(task = sched_useTask(root))) |
|
return NULL; |
|
|
|
task->task_func = func; |
|
TASK_TYPE(task) = taskTHREAD; |
|
TASK_ROOT(task) = root; |
|
|
|
TASK_ARG(task) = arg; |
|
TASK_FLAG(task) = detach ? PTHREAD_CREATE_DETACHED : PTHREAD_CREATE_JOINABLE; |
|
|
|
TASK_DATA(task) = opt_data; |
|
TASK_DATLEN(task) = opt_dlen; |
|
|
|
pthread_attr_init(&attr); |
|
pthread_attr_setdetachstate(&attr, TASK_FLAG(task)); |
|
if (root->root_hooks.hook_add.thread) |
|
ptr = root->root_hooks.hook_add.thread(task, &attr); |
|
else |
|
ptr = NULL; |
|
pthread_attr_destroy(&attr); |
|
|
|
if (!ptr) { |
|
pthread_mutex_lock(&root->root_mtx[taskTHREAD]); |
|
TAILQ_INSERT_TAIL(&root->root_thread, TASK_ID(task), task_node); |
|
pthread_mutex_unlock(&root->root_mtx[taskTHREAD]); |
|
} else |
|
task = sched_unuseTask(task); |
|
|
|
return task; |
|
} |
|
|