version 1.15.2.1, 2012/08/21 13:15:49
|
version 1.17.2.1, 2013/05/26 20:14:02
|
Line 12 terms:
|
Line 12 terms:
|
All of the documentation and software included in the ELWIX and AITNET |
All of the documentation and software included in the ELWIX and AITNET |
Releases is copyrighted by ELWIX - Sofia/Bulgaria <info@elwix.org> |
Releases is copyrighted by ELWIX - Sofia/Bulgaria <info@elwix.org> |
|
|
Copyright 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012 | Copyright 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012, 2013 |
by Michael Pounov <misho@elwix.org>. All rights reserved. |
by Michael Pounov <misho@elwix.org>. All rights reserved. |
|
|
Redistribution and use in source and binary forms, with or without |
Redistribution and use in source and binary forms, with or without |
Line 241 struct sched_RootTask {
|
Line 241 struct sched_RootTask {
|
#define ROOT_RETURN(x) (x)->root_ret |
#define ROOT_RETURN(x) (x)->root_ret |
|
|
|
|
inline int sched_GetErrno(); | int sched_GetErrno(); |
inline const char *sched_GetError(); | const char *sched_GetError(); |
|
|
|
|
/* |
/* |
Line 274 int schedRegisterHooks(sched_root_task_t * __restrict
|
Line 274 int schedRegisterHooks(sched_root_task_t * __restrict
|
* @root = root task |
* @root = root task |
* return: NULL error or !=NULL prepared task |
* return: NULL error or !=NULL prepared task |
*/ |
*/ |
inline sched_task_t *sched_useTask(sched_root_task_t * __restrict root); | sched_task_t *sched_useTask(sched_root_task_t * __restrict root); |
/* |
/* |
* sched_unuseTask() - Unlock and put task to unuse queue |
* sched_unuseTask() - Unlock and put task to unuse queue |
* |
* |
* @task = task |
* @task = task |
* return: always is NULL |
* return: always is NULL |
*/ |
*/ |
inline sched_task_t *sched_unuseTask(sched_task_t * __restrict task); | sched_task_t *sched_unuseTask(sched_task_t * __restrict task); |
/* |
/* |
* schedPolling() - Polling timeout period if no timer task is present |
* schedPolling() - Polling timeout period if no timer task is present |
* |
* |
Line 290 inline sched_task_t *sched_unuseTask(sched_task_t * __
|
Line 290 inline sched_task_t *sched_unuseTask(sched_task_t * __
|
* @tsold = old timeout polling if !=NULL |
* @tsold = old timeout polling if !=NULL |
* return: -1 error or 0 ok |
* return: -1 error or 0 ok |
*/ |
*/ |
inline int schedPolling(sched_root_task_t * __restrict root, | int schedPolling(sched_root_task_t * __restrict root, |
struct timespec * __restrict ts, struct timespec * __restrict tsold); |
struct timespec * __restrict ts, struct timespec * __restrict tsold); |
/* |
/* |
* schedTermCondition() - Activate hook for scheduler condition kill |
* schedTermCondition() - Activate hook for scheduler condition kill |
Line 299 inline int schedPolling(sched_root_task_t * __restrict
|
Line 299 inline int schedPolling(sched_root_task_t * __restrict
|
* @condValue = condition value, kill schedRun() if condValue == killState |
* @condValue = condition value, kill schedRun() if condValue == killState |
* return: -1 error or 0 ok |
* return: -1 error or 0 ok |
*/ |
*/ |
inline int schedTermCondition(sched_root_task_t * __restrict root, intptr_t condValue); | int schedTermCondition(sched_root_task_t * __restrict root, intptr_t condValue); |
/* |
/* |
* schedCall() - Call task execution function |
* schedCall() - Call task execution function |
* |
* |
* @task = current task |
* @task = current task |
* return: !=NULL error or =NULL ok |
* return: !=NULL error or =NULL ok |
*/ |
*/ |
inline void *schedCall(sched_task_t * __restrict task); | void *schedCall(sched_task_t * __restrict task); |
/* |
/* |
* schedFetch() - Fetch ready task |
* schedFetch() - Fetch ready task |
* |
* |
* @root = root task |
* @root = root task |
* return: =NULL error or !=NULL ready task |
* return: =NULL error or !=NULL ready task |
*/ |
*/ |
inline void *schedFetch(sched_root_task_t * __restrict root); | void *schedFetch(sched_root_task_t * __restrict root); |
/* |
/* |
* schedRun() - Scheduler *run loop* |
* schedRun() - Scheduler *run loop* |
* |
* |
Line 461 sched_task_t *schedAIO(sched_root_task_t * __restrict
|
Line 461 sched_task_t *schedAIO(sched_root_task_t * __restrict
|
* @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
* return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
*/ |
*/ |
inline sched_task_t *schedAIORead(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedAIORead(sched_root_task_t * __restrict root, sched_task_func_t func, |
void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
/* |
/* |
* schedAIOWrite() - Add AIO write task to scheduler queue |
* schedAIOWrite() - Add AIO write task to scheduler queue |
Line 475 inline sched_task_t *schedAIORead(sched_root_task_t *
|
Line 475 inline sched_task_t *schedAIORead(sched_root_task_t *
|
* @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
* return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
*/ |
*/ |
inline sched_task_t *schedAIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedAIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, |
void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
void *arg, int fd, void *buffer, size_t buflen, off_t offset); |
|
|
/* |
/* |
Line 503 sched_task_t *schedLIO(sched_root_task_t * __restrict
|
Line 503 sched_task_t *schedLIO(sched_root_task_t * __restrict
|
* @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
* return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
*/ |
*/ |
inline sched_task_t *schedLIORead(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedLIORead(sched_root_task_t * __restrict root, sched_task_func_t func, |
void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
/* |
/* |
* schedLIOWrite() - Add list of AIO write tasks to scheduler queue |
* schedLIOWrite() - Add list of AIO write tasks to scheduler queue |
Line 517 inline sched_task_t *schedLIORead(sched_root_task_t *
|
Line 517 inline sched_task_t *schedLIORead(sched_root_task_t *
|
* @offset = Offset from start of file, if =-1 from current position |
* @offset = Offset from start of file, if =-1 from current position |
* return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
*/ |
*/ |
inline sched_task_t *schedLIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, | sched_task_t *schedLIOWrite(sched_root_task_t * __restrict root, sched_task_func_t func, |
void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
void *arg, int fd, struct iovec *bufs, size_t nbufs, off_t offset); |
#endif /* EVFILT_LIO */ |
#endif /* EVFILT_LIO */ |
|
|
Line 638 sched_task_t *schedCallOnce(sched_root_task_t * __rest
|
Line 638 sched_task_t *schedCallOnce(sched_root_task_t * __rest
|
* @func = task execution function |
* @func = task execution function |
* @arg = 1st func argument |
* @arg = 1st func argument |
* @detach = Detach thread from scheduler, if !=0 |
* @detach = Detach thread from scheduler, if !=0 |
|
* @ss = stack size |
* @opt_data = Optional data |
* @opt_data = Optional data |
* @opt_dlen = Optional data length |
* @opt_dlen = Optional data length |
* return: NULL error or !=NULL new queued task |
* return: NULL error or !=NULL new queued task |
*/ |
*/ |
sched_task_t *schedThread(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, |
sched_task_t *schedThread(sched_root_task_t * __restrict root, sched_task_func_t func, void *arg, |
int detach, void *opt_data, size_t opt_dlen); | int detach, size_t ss, void *opt_data, size_t opt_dlen); |
#define schedThreadSelf(x) schedThread(TASK_ROOT((x)), TASK_FUNC((x)), TASK_ARG((x)), \ |
#define schedThreadSelf(x) schedThread(TASK_ROOT((x)), TASK_FUNC((x)), TASK_ARG((x)), \ |
TASK_VAL((x)), TASK_DATA((x)), TASK_DATLEN((x))) | (TASK_FLAG((x)) & 0x1), (size_t) (TASK_FLAG((x)) >> 1), TASK_DATA((x)), TASK_DATLEN((x))) |
/* |
/* |
* taskExit() - Exit routine for scheduler task | * sched_taskExit() - Exit routine for scheduler task, explicit required for thread tasks |
* |
* |
|
* @task = current task |
|
* @retcode = return code |
|
* return: return code |
|
*/ |
|
void *sched_taskExit(sched_task_t *task, intptr_t retcode); |
|
/* |
|
* taskExit() - Exit helper for scheduler task |
|
* |
* @t = current executed task |
* @t = current executed task |
* @x = exit value for task |
* @x = exit value for task |
* return: none |
* return: none |
*/ |
*/ |
#define taskExit(t, x) do { assert((t) && TASK_ROOT(t)); \ | #define taskExit(t, x) return sched_taskExit((t), (intptr_t) (x)) |
if (TASK_ROOT(t)->root_hooks.hook_exec.exit) \ | |
TASK_ROOT(t)->root_hooks.hook_exec.exit((t), \ | |
(void*) (x)); \ | |
TASK_ROOT(t)->root_ret = (void*) (x); \ | |
if (TASK_TYPE(t) == taskTHREAD) { \ | |
sched_unuseTask(t); \ | |
pthread_exit((void*) (x)); \ | |
} else \ | |
return ((void*) (x)); \ | |
} while (0) | |
#define taskKill(t, s) do { assert((t) && TASK_ROOT(t)); \ | |
if (TASK_TYPE(t) == taskTHREAD) { \ | |
pthread_t _tid = (pthread_t) TASK_VAL((t)); \ | |
sched_unuseTask(t); \ | |
pthread_kill(_tid, (s)); \ | |
} else \ | |
schedCancel((t)); \ | |
} while (0) | |
|
|
|
|
#endif |
#endif |