|
version 1.15.2.1, 2012/08/21 13:15:49
|
version 1.18, 2013/05/30 09:13:52
|
|
Line 12 terms:
|
Line 12 terms:
|
| All of the documentation and software included in the ELWIX and AITNET |
All of the documentation and software included in the ELWIX and AITNET |
| Releases is copyrighted by ELWIX - Sofia/Bulgaria <info@elwix.org> |
Releases is copyrighted by ELWIX - Sofia/Bulgaria <info@elwix.org> |
| |
|
| Copyright 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012 | Copyright 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012, 2013 |
| by Michael Pounov <misho@elwix.org>. All rights reserved. |
by Michael Pounov <misho@elwix.org>. All rights reserved. |
| |
|
| Redistribution and use in source and binary forms, with or without |
Redistribution and use in source and binary forms, with or without |
|
Line 241 struct sched_RootTask {
|
Line 241 struct sched_RootTask {
|
| #define ROOT_RETURN(x) (x)->root_ret |
#define ROOT_RETURN(x) (x)->root_ret |
| |
|
| |
|
| inline int sched_GetErrno(); | int sched_GetErrno(); |
| inline const char *sched_GetError(); | const char *sched_GetError(); |
| |
|
| |
|
| /* |
/* |
|
Line 274 int schedRegisterHooks(sched_root_task_t * __restrict
|
Line 274 int schedRegisterHooks(sched_root_task_t * __restrict
|
| * @root = root task |
* @root = root task |
| * return: NULL error or !=NULL prepared task |
* return: NULL error or !=NULL prepared task |
| */ |
*/ |
| inline sched_task_t *sched_useTask(sched_root_task_t * __restrict root); | sched_task_t *sched_useTask(sched_root_task_t * __restrict root); |
| /* |
/* |
| * sched_unuseTask() - Unlock and put task to unuse queue |
* sched_unuseTask() - Unlock and put task to unuse queue |
| * |
* |
| * @task = task |
* @task = task |
| * return: always is NULL |
* return: always is NULL |
| */ |
*/ |
| inline sched_task_t *sched_unuseTask(sched_task_t * __restrict task); | sched_task_t *sched_unuseTask(sched_task_t * __restrict task); |
| /* |
/* |
| * schedPolling() - Polling timeout period if no timer task is present |
* schedPolling() - Polling timeout period if no timer task is present |
| * |
* |
|
Line 290 inline sched_task_t *sched_unuseTask(sched_task_t * __
|
Line 290 inline sched_task_t *sched_unuseTask(sched_task_t * __
|
| * @tsold = old timeout polling if !=NULL |
* @tsold = old timeout polling if !=NULL |
| * return: -1 error or 0 ok |
* return: -1 error or 0 ok |
| */ |
*/ |
| inline int schedPolling(sched_root_task_t * __restrict root, | int schedPolling(sched_root_task_t * __restrict root, |
| struct timespec * __restrict ts, struct timespec * __restrict tsold); |
struct timespec * __restrict ts, struct timespec * __restrict tsold); |
| /* |
/* |
| * schedTermCondition() - Activate hook for scheduler condition kill |
* schedTermCondition() - Activate hook for scheduler condition kill |
|
Line 299 inline int schedPolling(sched_root_task_t * __restrict
|
Line 299 inline int schedPolling(sched_root_task_t * __restrict
|
| * @condValue = condition value, kill schedRun() if condValue == killState |
* @condValue = condition value, kill schedRun() if condValue == killState |
| * return: -1 error or 0 ok |
* return: -1 error or 0 ok |
| */ |
*/ |
| inline int schedTermCondition(sched_root_task_t * __restrict root, intptr_t condValue); | int schedTermCondition(sched_root_task_t * __restrict root, intptr_t condValue); |
| /* |
/* |
| * schedCall() - Call task execution function |
* schedCall() - Call task execution function |
| * |
* |
| * @task = current task |
* @task = current task |
| * return: !=NULL error or =NULL ok |
* return: !=NULL error or =NULL ok |
| */ |
*/ |
| inline void *schedCall(sched_task_t * __restrict task); | void *schedCall(sched_task_t * __restrict task); |
| /* |
/* |
| * schedFetch() - Fetch ready task |
* schedFetch() - Fetch ready task |
| * |
* |
| * @root = root task |
* @root = root task |
| * return: =NULL error or !=NULL ready task |
* return: =NULL error or !=NULL ready task |
| */ |
*/ |
| inline void *schedFetch(sched_root_task_t * __restrict root); | void *schedFetch(sched_root_task_t * __restrict root); |
| /* |
/* |
| * schedRun() - Scheduler *run loop* |
* schedRun() - Scheduler *run loop* |
| * |
* |
|
Line 461 sched_task_t *schedAIO(sched_root_task_t * __restrict
|
Line 461 sched_task_t *schedAIO(sched_root_task_t * __restrict
|
| * @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
| * return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
| */ |
*/ |
| inline sched_task_t *schedAIORead(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedAIORead(sched_root_task_t * __restrict root, sched_task_func_t func, |
| void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
| /* |
/* |
| * schedAIOWrite() - Add AIO write task to scheduler queue |
* schedAIOWrite() - Add AIO write task to scheduler queue |
|
Line 475 inline sched_task_t *schedAIORead(sched_root_task_t *
|
Line 475 inline sched_task_t *schedAIORead(sched_root_task_t *
|
| * @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
| * return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
| */ |
*/ |
| inline sched_task_t *schedAIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedAIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, |
| void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
| |
|
| /* |
/* |
|
Line 503 sched_task_t *schedLIO(sched_root_task_t * __restrict
|
Line 503 sched_task_t *schedLIO(sched_root_task_t * __restrict
|
| * @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
| * return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
| */ |
*/ |
| inline sched_task_t *schedLIORead(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedLIORead(sched_root_task_t * __restrict root, sched_task_func_t func, |
| void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
| /* |
/* |
| * schedLIOWrite() - Add list of AIO write tasks to scheduler queue |
* schedLIOWrite() - Add list of AIO write tasks to scheduler queue |
|
Line 517 inline sched_task_t *schedLIORead(sched_root_task_t *
|
Line 517 inline sched_task_t *schedLIORead(sched_root_task_t *
|
| * @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
| * return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
| */ |
*/ |
| inline sched_task_t *schedLIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedLIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, |
| void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
| #endif /* EVFILT_LIO */ |
#endif /* EVFILT_LIO */ |
| |
|
|
Line 638 sched_task_t *schedCallOnce(sched_root_task_t * __rest
|
Line 638 sched_task_t *schedCallOnce(sched_root_task_t * __rest
|
| * @func = task execution function |
* @func = task execution function |
| * @arg = 1st func argument |
* @arg = 1st func argument |
| * @detach = Detach thread from scheduler, if !=0 |
* @detach = Detach thread from scheduler, if !=0 |
| |
* @ss = stack size |
| * @opt_data = Optional data |
* @opt_data = Optional data |
| * @opt_dlen = Optional data length |
* @opt_dlen = Optional data length |
| * return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
| */ |
*/ |
| sched_task_t *schedThread(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, |
sched_task_t *schedThread(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, |
| int detach, void *opt_data, size_t opt_dlen); | int detach, size_t ss, void *opt_data, size_t opt_dlen); |
| #define schedThreadSelf(x) schedThread(TASK_ROOT((x)), TASK_FUNC((x)), TASK_ARG((x)), \ |
#define schedThreadSelf(x) schedThread(TASK_ROOT((x)), TASK_FUNC((x)), TASK_ARG((x)), \ |
| TASK_VAL((x)), TASK_DATA((x)), TASK_DATLEN((x))) | (TASK_FLAG((x)) & 0x1), (size_t) (TASK_FLAG((x)) >> 1), TASK_DATA((x)), TASK_DATLEN((x))) |
| /* |
/* |
| * taskExit() - Exit routine for scheduler task | * sched_taskExit() - Exit routine for scheduler task, explicit required for thread tasks |
| * |
* |
| |
* @task = current task |
| |
* @retcode = return code |
| |
* return: return code |
| |
*/ |
| |
void *sched_taskExit(sched_task_t *task, intptr_t retcode); |
| |
/* |
| |
* taskExit() - Exit helper for scheduler task |
| |
* |
| * @t = current executed task |
* @t = current executed task |
| * @x = exit value for task |
* @x = exit value for task |
| * return: none |
* return: none |
| */ |
*/ |
| #define taskExit(t, x) do { assert((t) && TASK_ROOT(t)); \ | #define taskExit(t, x) return sched_taskExit((t), (intptr_t) (x)) |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exit) \ | |
| TASK_ROOT(t)->root_hooks.hook_exec.exit((t), \ | |
| (void*) (x)); \ | |
| TASK_ROOT(t)->root_ret = (void*) (x); \ | |
| if (TASK_TYPE(t) == taskTHREAD) { \ | |
| sched_unuseTask(t); \ | |
| pthread_exit((void*) (x)); \ | |
| } else \ | |
| return ((void*) (x)); \ | |
| } while (0) | |
| #define taskKill(t, s) do { assert((t) && TASK_ROOT(t)); \ | |
| if (TASK_TYPE(t) == taskTHREAD) { \ | |
| pthread_t _tid = (pthread_t) TASK_VAL((t)); \ | |
| sched_unuseTask(t); \ | |
| pthread_kill(_tid, (s)); \ | |
| } else \ | |
| schedCancel((t)); \ | |
| } while (0) | |
| |
|
| |
|
| #endif |
#endif |