version 1.13.2.1, 2012/08/21 11:07:16
|
version 1.27.2.3, 2014/05/21 22:09:01
|
Line 12 terms:
|
Line 12 terms:
|
All of the documentation and software included in the ELWIX and AITNET |
All of the documentation and software included in the ELWIX and AITNET |
Releases is copyrighted by ELWIX - Sofia/Bulgaria <info@elwix.org> |
Releases is copyrighted by ELWIX - Sofia/Bulgaria <info@elwix.org> |
|
|
Copyright 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012 | Copyright 2004 - 2014 |
by Michael Pounov <misho@elwix.org>. All rights reserved. |
by Michael Pounov <misho@elwix.org>. All rights reserved. |
|
|
Redistribution and use in source and binary forms, with or without |
Redistribution and use in source and binary forms, with or without |
Line 62 sched_hook_init(void *root, void *arg __unused)
|
Line 62 sched_hook_init(void *root, void *arg __unused)
|
if (!r) |
if (!r) |
return (void*) -1; |
return (void*) -1; |
|
|
|
#if SUP_ENABLE == KQ_ENABLE |
r->root_kq = kqueue(); |
r->root_kq = kqueue(); |
if (r->root_kq == -1) { |
if (r->root_kq == -1) { |
LOGERR; |
LOGERR; |
return (void*) -1; |
return (void*) -1; |
} |
} |
|
#else |
|
r->root_kq ^= r->root_kq; |
|
FD_ZERO(&r->root_fds[0]); |
|
FD_ZERO(&r->root_fds[1]); |
|
#endif |
|
|
return NULL; |
return NULL; |
} |
} |
Line 86 sched_hook_fini(void *root, void *arg __unused)
|
Line 92 sched_hook_fini(void *root, void *arg __unused)
|
if (!r) |
if (!r) |
return (void*) -1; |
return (void*) -1; |
|
|
|
#if SUP_ENABLE == KQ_ENABLE |
if (r->root_kq > 2) { |
if (r->root_kq > 2) { |
close(r->root_kq); |
close(r->root_kq); |
r->root_kq = 0; |
r->root_kq = 0; |
} |
} |
|
#else |
|
FD_ZERO(&r->root_fds[1]); |
|
FD_ZERO(&r->root_fds[0]); |
|
r->root_kq ^= r->root_kq; |
|
#endif |
|
|
return NULL; |
return NULL; |
} |
} |
Line 105 void *
|
Line 117 void *
|
sched_hook_cancel(void *task, void *arg __unused) |
sched_hook_cancel(void *task, void *arg __unused) |
{ |
{ |
sched_task_t *t = task; |
sched_task_t *t = task; |
|
#if SUP_ENABLE == KQ_ENABLE |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
|
#else |
|
sched_root_task_t *r = NULL; |
|
register int i; |
|
#endif |
#ifdef AIO_SUPPORT |
#ifdef AIO_SUPPORT |
struct aiocb *acb; |
struct aiocb *acb; |
#ifdef EVFILT_LIO |
#ifdef EVFILT_LIO |
|
register int i = 0; |
struct aiocb **acbs; |
struct aiocb **acbs; |
register int i; |
|
#endif /* EVFILT_LIO */ |
#endif /* EVFILT_LIO */ |
#endif /* AIO_SUPPORT */ |
#endif /* AIO_SUPPORT */ |
|
|
if (!t || !TASK_ROOT(t)) |
if (!t || !TASK_ROOT(t)) |
return (void*) -1; |
return (void*) -1; |
|
#if SUP_ENABLE != KQ_ENABLE |
|
r = TASK_ROOT(t); |
|
#endif |
|
|
switch (TASK_TYPE(t)) { |
switch (TASK_TYPE(t)) { |
case taskREAD: |
case taskREAD: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
|
#else |
|
FD_CLR(TASK_FD(t), &r->root_fds[0]); |
|
|
|
/* optimize select */ |
|
for (i = r->root_kq - 1; i > 2; i--) |
|
if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) |
|
break; |
|
if (i > 2) |
|
r->root_kq = i + 1; |
|
#endif |
break; |
break; |
case taskWRITE: |
case taskWRITE: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
|
#else |
|
FD_CLR(TASK_FD(t), &r->root_fds[1]); |
|
|
|
/* optimize select */ |
|
for (i = r->root_kq - 1; i > 2; i--) |
|
if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) |
|
break; |
|
if (i > 2) |
|
r->root_kq = i + 1; |
|
#endif |
break; |
break; |
case taskALARM: |
case taskALARM: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, |
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, |
0, 0, (intptr_t) TASK_DATA(t)); |
0, 0, (intptr_t) TASK_DATA(t)); |
Line 141 sched_hook_cancel(void *task, void *arg __unused)
|
Line 184 sched_hook_cancel(void *task, void *arg __unused)
|
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, |
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, |
0, 0, (void*) TASK_DATA(t)); |
0, 0, (void*) TASK_DATA(t)); |
#endif |
#endif |
|
#endif |
break; |
break; |
case taskNODE: |
case taskNODE: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
|
#endif |
break; |
break; |
case taskPROC: |
case taskPROC: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
#endif |
#endif |
|
#endif |
break; |
break; |
case taskSIGNAL: |
case taskSIGNAL: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
#endif |
#endif |
|
/* restore signal */ |
|
signal(TASK_VAL(t), SIG_DFL); |
|
#endif |
break; |
break; |
#ifdef AIO_SUPPORT |
#ifdef AIO_SUPPORT |
case taskAIO: |
case taskAIO: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
#else |
#else |
Line 177 sched_hook_cancel(void *task, void *arg __unused)
|
Line 230 sched_hook_cancel(void *task, void *arg __unused)
|
free(acb); |
free(acb); |
TASK_VAL(t) = 0; |
TASK_VAL(t) = 0; |
} |
} |
|
#endif |
break; |
break; |
#ifdef EVFILT_LIO |
#ifdef EVFILT_LIO |
case taskLIO: |
case taskLIO: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
#else |
#else |
Line 195 sched_hook_cancel(void *task, void *arg __unused)
|
Line 250 sched_hook_cancel(void *task, void *arg __unused)
|
free(acbs); |
free(acbs); |
TASK_VAL(t) = 0; |
TASK_VAL(t) = 0; |
} |
} |
|
#endif |
break; |
break; |
#endif /* EVFILT_LIO */ |
#endif /* EVFILT_LIO */ |
#endif /* AIO_SUPPORT */ |
#endif /* AIO_SUPPORT */ |
#ifdef EVFILT_USER |
#ifdef EVFILT_USER |
case taskUSER: |
case taskUSER: |
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
#endif |
#endif |
break; |
|
#endif |
#endif |
|
break; |
|
#endif /* EVFILT_USER */ |
case taskTHREAD: |
case taskTHREAD: |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_cancel((pthread_t) TASK_VAL(t)); | if (TASK_VAL(t)) |
| pthread_cancel((pthread_t) TASK_VAL(t)); |
#endif |
#endif |
|
return NULL; |
|
#if defined(HAVE_TIMER_CREATE) && defined(HAVE_TIMER_SETTIME) && defined(HAVE_TIMER_DELETE) |
|
case taskRTC: |
|
timer_delete((timer_t) TASK_FLAG(t)); |
|
schedCancel((sched_task_t*) TASK_RET(t)); |
|
return NULL; |
|
#endif /* HAVE_TIMER_CREATE */ |
default: |
default: |
return NULL; |
return NULL; |
} |
} |
|
|
|
#if SUP_ENABLE == KQ_ENABLE |
kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); |
kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); |
|
#endif |
return NULL; |
return NULL; |
} |
} |
|
|
|
#ifdef HAVE_LIBPTHREAD |
/* |
/* |
|
* sched_hook_thread() - Default THREAD hook |
|
* |
|
* @task = current task |
|
* @arg = pthread attributes |
|
* return: <0 errors and 0 ok |
|
*/ |
|
void * |
|
sched_hook_thread(void *task, void *arg) |
|
{ |
|
sched_task_t *t = task; |
|
pthread_t tid; |
|
sigset_t s, o; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
sigfillset(&s); |
|
pthread_sigmask(SIG_BLOCK, &s, &o); |
|
if ((errno = pthread_create(&tid, (pthread_attr_t*) arg, |
|
(void *(*)(void*)) _sched_threadWrapper, t))) { |
|
LOGERR; |
|
pthread_sigmask(SIG_SETMASK, &o, NULL); |
|
return (void*) -1; |
|
} else |
|
TASK_VAL(t) = (u_long) tid; |
|
|
|
if (!TASK_ISLOCKED(t)) |
|
TASK_LOCK(t); |
|
|
|
pthread_sigmask(SIG_SETMASK, &o, NULL); |
|
return NULL; |
|
} |
|
#endif |
|
|
|
/* |
* sched_hook_read() - Default READ hook |
* sched_hook_read() - Default READ hook |
* |
* |
* @task = current task |
* @task = current task |
Line 230 void *
|
Line 334 void *
|
sched_hook_read(void *task, void *arg __unused) |
sched_hook_read(void *task, void *arg __unused) |
{ |
{ |
sched_task_t *t = task; |
sched_task_t *t = task; |
|
#if SUP_ENABLE == KQ_ENABLE |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
|
#else |
|
sched_root_task_t *r = NULL; |
|
#endif |
|
|
if (!t || !TASK_ROOT(t)) |
if (!t || !TASK_ROOT(t)) |
return (void*) -1; |
return (void*) -1; |
|
#if SUP_ENABLE != KQ_ENABLE |
|
r = TASK_ROOT(t); |
|
#endif |
|
|
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
Line 248 sched_hook_read(void *task, void *arg __unused)
|
Line 360 sched_hook_read(void *task, void *arg __unused)
|
LOGERR; |
LOGERR; |
return (void*) -1; |
return (void*) -1; |
} |
} |
|
#else |
|
FD_SET(TASK_FD(t), &r->root_fds[0]); |
|
if (TASK_FD(t) >= r->root_kq) |
|
r->root_kq = TASK_FD(t) + 1; |
|
#endif |
|
|
return NULL; |
return NULL; |
} |
} |
Line 263 void *
|
Line 380 void *
|
sched_hook_write(void *task, void *arg __unused) |
sched_hook_write(void *task, void *arg __unused) |
{ |
{ |
sched_task_t *t = task; |
sched_task_t *t = task; |
|
#if SUP_ENABLE == KQ_ENABLE |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
|
#else |
|
sched_root_task_t *r = NULL; |
|
#endif |
|
|
if (!t || !TASK_ROOT(t)) |
if (!t || !TASK_ROOT(t)) |
return (void*) -1; |
return (void*) -1; |
|
#if SUP_ENABLE != KQ_ENABLE |
|
r = TASK_ROOT(t); |
|
#endif |
|
|
|
#if SUP_ENABLE == KQ_ENABLE |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
Line 281 sched_hook_write(void *task, void *arg __unused)
|
Line 406 sched_hook_write(void *task, void *arg __unused)
|
LOGERR; |
LOGERR; |
return (void*) -1; |
return (void*) -1; |
} |
} |
|
#else |
|
FD_SET(TASK_FD(t), &r->root_fds[1]); |
|
if (TASK_FD(t) >= r->root_kq) |
|
r->root_kq = TASK_FD(t) + 1; |
|
#endif |
|
|
return NULL; |
return NULL; |
} |
} |
Line 295 sched_hook_write(void *task, void *arg __unused)
|
Line 425 sched_hook_write(void *task, void *arg __unused)
|
void * |
void * |
sched_hook_alarm(void *task, void *arg __unused) |
sched_hook_alarm(void *task, void *arg __unused) |
{ |
{ |
|
#if SUP_ENABLE == KQ_ENABLE |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
Line 303 sched_hook_alarm(void *task, void *arg __unused)
|
Line 434 sched_hook_alarm(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
|
|
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, | EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_CLEAR, 0, |
t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, |
t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, |
(intptr_t) TASK_DATA(t)); |
(intptr_t) TASK_DATA(t)); |
#else |
#else |
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, | EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_CLEAR, 0, |
t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, |
t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, |
(void*) TASK_DATA(t)); |
(void*) TASK_DATA(t)); |
#endif |
#endif |
Line 319 sched_hook_alarm(void *task, void *arg __unused)
|
Line 450 sched_hook_alarm(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
} |
} |
|
|
|
#endif |
return NULL; |
return NULL; |
} |
} |
|
|
Line 332 sched_hook_alarm(void *task, void *arg __unused)
|
Line 464 sched_hook_alarm(void *task, void *arg __unused)
|
void * |
void * |
sched_hook_node(void *task, void *arg __unused) |
sched_hook_node(void *task, void *arg __unused) |
{ |
{ |
|
#if SUP_ENABLE == KQ_ENABLE |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
Line 356 sched_hook_node(void *task, void *arg __unused)
|
Line 489 sched_hook_node(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
} |
} |
|
|
|
#endif |
return NULL; |
return NULL; |
} |
} |
|
|
Line 369 sched_hook_node(void *task, void *arg __unused)
|
Line 503 sched_hook_node(void *task, void *arg __unused)
|
void * |
void * |
sched_hook_proc(void *task, void *arg __unused) |
sched_hook_proc(void *task, void *arg __unused) |
{ |
{ |
|
#if SUP_ENABLE == KQ_ENABLE |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
Line 391 sched_hook_proc(void *task, void *arg __unused)
|
Line 526 sched_hook_proc(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
} |
} |
|
|
|
#endif |
return NULL; |
return NULL; |
} |
} |
|
|
Line 404 sched_hook_proc(void *task, void *arg __unused)
|
Line 540 sched_hook_proc(void *task, void *arg __unused)
|
void * |
void * |
sched_hook_signal(void *task, void *arg __unused) |
sched_hook_signal(void *task, void *arg __unused) |
{ |
{ |
|
#if SUP_ENABLE == KQ_ENABLE |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
Line 411 sched_hook_signal(void *task, void *arg __unused)
|
Line 548 sched_hook_signal(void *task, void *arg __unused)
|
if (!t || !TASK_ROOT(t)) |
if (!t || !TASK_ROOT(t)) |
return (void*) -1; |
return (void*) -1; |
|
|
|
/* ignore signal */ |
|
signal(TASK_VAL(t), SIG_IGN); |
|
|
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (intptr_t) TASK_VAL(t)); | EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_VAL(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (void*) TASK_VAL(t)); | EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_VAL(t)); |
#endif |
#endif |
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
Line 423 sched_hook_signal(void *task, void *arg __unused)
|
Line 563 sched_hook_signal(void *task, void *arg __unused)
|
LOGERR; |
LOGERR; |
return (void*) -1; |
return (void*) -1; |
} |
} |
|
#else |
|
#if 0 |
|
sched_task_t *t = task; |
|
struct sigaction sa; |
|
|
|
memset(&sa, 0, sizeof sa); |
|
sigemptyset(&sa.sa_mask); |
|
sa.sa_handler = _sched_sigHandler; |
|
sa.sa_flags = SA_RESETHAND | SA_RESTART; |
|
|
|
if (sigaction(TASK_VAL(t), &sa, NULL) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} |
|
#endif /* 0 */ |
|
#endif |
return NULL; |
return NULL; |
} |
} |
|
|
Line 438 sched_hook_signal(void *task, void *arg __unused)
|
Line 596 sched_hook_signal(void *task, void *arg __unused)
|
void * |
void * |
sched_hook_user(void *task, void *arg __unused) |
sched_hook_user(void *task, void *arg __unused) |
{ |
{ |
|
#if SUP_ENABLE == KQ_ENABLE |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
Line 460 sched_hook_user(void *task, void *arg __unused)
|
Line 619 sched_hook_user(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
} |
} |
|
|
|
#endif |
return NULL; |
return NULL; |
} |
} |
#endif |
#endif |
Line 477 sched_hook_fetch(void *root, void *arg __unused)
|
Line 637 sched_hook_fetch(void *root, void *arg __unused)
|
sched_root_task_t *r = root; |
sched_root_task_t *r = root; |
sched_task_t *task, *tmp; |
sched_task_t *task, *tmp; |
struct timespec now, m, mtmp; |
struct timespec now, m, mtmp; |
struct timespec *timeout; | #if SUP_ENABLE == KQ_ENABLE |
struct kevent evt[1], res[KQ_EVENTS]; |
struct kevent evt[1], res[KQ_EVENTS]; |
|
struct timespec *timeout; |
|
#else |
|
struct timeval *timeout, tv; |
|
fd_set rfd, wfd, xfd; |
|
#endif |
register int i, flg; |
register int i, flg; |
int en; |
int en; |
#ifdef AIO_SUPPORT |
#ifdef AIO_SUPPORT |
Line 553 sched_hook_fetch(void *root, void *arg __unused)
|
Line 718 sched_hook_fetch(void *root, void *arg __unused)
|
/* set wait INFTIM */ |
/* set wait INFTIM */ |
sched_timespecinf(&r->root_wait); |
sched_timespecinf(&r->root_wait); |
} |
} |
#else | #else /* ! TIMER_WITHOUT_SORT */ |
if (!TAILQ_FIRST(&r->root_task) && (task = TAILQ_FIRST(&r->root_timer))) { |
if (!TAILQ_FIRST(&r->root_task) && (task = TAILQ_FIRST(&r->root_timer))) { |
clock_gettime(CLOCK_MONOTONIC, &now); |
clock_gettime(CLOCK_MONOTONIC, &now); |
|
|
Line 564 sched_hook_fetch(void *root, void *arg __unused)
|
Line 729 sched_hook_fetch(void *root, void *arg __unused)
|
/* set wait INFTIM */ |
/* set wait INFTIM */ |
sched_timespecinf(&r->root_wait); |
sched_timespecinf(&r->root_wait); |
} |
} |
#endif | #endif /* TIMER_WITHOUT_SORT */ |
/* if present member of task, set NOWAIT */ |
/* if present member of task, set NOWAIT */ |
if (TAILQ_FIRST(&r->root_task)) |
if (TAILQ_FIRST(&r->root_task)) |
sched_timespecclear(&r->root_wait); |
sched_timespecclear(&r->root_wait); |
|
|
if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) | if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) { |
| #if SUP_ENABLE == KQ_ENABLE |
timeout = &r->root_wait; |
timeout = &r->root_wait; |
else if (sched_timespecisinf(&r->root_poll)) | #else |
| sched_timespec2val(&r->root_wait, &tv); |
| timeout = &tv; |
| #endif /* KQ_SUPPORT */ |
| } else if (sched_timespecisinf(&r->root_poll)) |
timeout = NULL; |
timeout = NULL; |
else | else { |
| #if SUP_ENABLE == KQ_ENABLE |
timeout = &r->root_poll; |
timeout = &r->root_poll; |
|
#else |
|
sched_timespec2val(&r->root_poll, &tv); |
|
timeout = &tv; |
|
#endif /* KQ_SUPPORT */ |
|
} |
|
|
|
#if SUP_ENABLE == KQ_ENABLE |
if ((en = kevent(r->root_kq, NULL, 0, res, KQ_EVENTS, timeout)) == -1) { |
if ((en = kevent(r->root_kq, NULL, 0, res, KQ_EVENTS, timeout)) == -1) { |
|
#else |
|
rfd = xfd = r->root_fds[0]; |
|
wfd = r->root_fds[1]; |
|
if ((en = select(r->root_kq, &rfd, &wfd, &xfd, timeout)) == -1) { |
|
#endif /* KQ_SUPPORT */ |
if (r->root_hooks.hook_exec.exception) { |
if (r->root_hooks.hook_exec.exception) { |
if (r->root_hooks.hook_exec.exception(r, NULL)) |
if (r->root_hooks.hook_exec.exception(r, NULL)) |
return NULL; |
return NULL; |
} else if (errno != EINTR) |
} else if (errno != EINTR) |
LOGERR; |
LOGERR; |
return NULL; | goto skip_event; |
} |
} |
|
|
|
/* kevent dispatcher */ |
now.tv_sec = now.tv_nsec = 0; |
now.tv_sec = now.tv_nsec = 0; |
/* Go and catch the cat into pipes ... */ |
/* Go and catch the cat into pipes ... */ |
|
#if SUP_ENABLE == KQ_ENABLE |
for (i = 0; i < en; i++) { |
for (i = 0; i < en; i++) { |
memcpy(evt, &res[i], sizeof evt); |
memcpy(evt, &res[i], sizeof evt); |
evt->flags = EV_DELETE; |
evt->flags = EV_DELETE; |
Line 596 sched_hook_fetch(void *root, void *arg __unused)
|
Line 781 sched_hook_fetch(void *root, void *arg __unused)
|
TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { |
TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
continue; |
continue; |
else | else { |
flg++; |
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = (u_long) res[i].fflags; |
|
} |
/* remove read handle */ |
/* remove read handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskREAD]); |
pthread_mutex_lock(&r->root_mtx[taskREAD]); |
Line 646 sched_hook_fetch(void *root, void *arg __unused)
|
Line 834 sched_hook_fetch(void *root, void *arg __unused)
|
TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { |
TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
continue; |
continue; |
else | else { |
flg++; |
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = (u_long) res[i].fflags; |
|
} |
/* remove write handle */ |
/* remove write handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskWRITE]); |
pthread_mutex_lock(&r->root_mtx[taskWRITE]); |
Line 696 sched_hook_fetch(void *root, void *arg __unused)
|
Line 887 sched_hook_fetch(void *root, void *arg __unused)
|
TAILQ_FOREACH_SAFE(task, &r->root_alarm, task_node, tmp) { |
TAILQ_FOREACH_SAFE(task, &r->root_alarm, task_node, tmp) { |
if ((uintptr_t) TASK_DATA(task) != ((uintptr_t) res[i].udata)) |
if ((uintptr_t) TASK_DATA(task) != ((uintptr_t) res[i].udata)) |
continue; |
continue; |
else | else { |
flg++; |
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = (u_long) res[i].fflags; |
|
} |
/* remove alarm handle */ |
/* remove alarm handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskALARM]); |
pthread_mutex_lock(&r->root_mtx[taskALARM]); |
Line 726 sched_hook_fetch(void *root, void *arg __unused)
|
Line 920 sched_hook_fetch(void *root, void *arg __unused)
|
continue; |
continue; |
else { |
else { |
flg++; |
flg++; |
TASK_DATA(task) = (void*) (uintptr_t) res[i].data; | TASK_RET(task) = res[i].data; |
TASK_DATLEN(task) = res[i].fflags; | TASK_FLAG(task) = (u_long) res[i].fflags; |
} |
} |
/* remove node handle */ |
/* remove node handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
Line 757 sched_hook_fetch(void *root, void *arg __unused)
|
Line 951 sched_hook_fetch(void *root, void *arg __unused)
|
continue; |
continue; |
else { |
else { |
flg++; |
flg++; |
TASK_DATA(task) = (void*) (uintptr_t) res[i].data; | TASK_RET(task) = res[i].data; |
TASK_DATLEN(task) = res[i].fflags; | TASK_FLAG(task) = (u_long) res[i].fflags; |
} |
} |
/* remove proc handle */ |
/* remove proc handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
Line 786 sched_hook_fetch(void *root, void *arg __unused)
|
Line 980 sched_hook_fetch(void *root, void *arg __unused)
|
TAILQ_FOREACH_SAFE(task, &r->root_signal, task_node, tmp) { |
TAILQ_FOREACH_SAFE(task, &r->root_signal, task_node, tmp) { |
if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) |
if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) |
continue; |
continue; |
else | else { |
flg++; |
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = (u_long) res[i].fflags; |
|
} |
/* remove signal handle */ |
/* remove signal handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskSIGNAL]); |
pthread_mutex_lock(&r->root_mtx[taskSIGNAL]); |
Line 816 sched_hook_fetch(void *root, void *arg __unused)
|
Line 1013 sched_hook_fetch(void *root, void *arg __unused)
|
acb = (struct aiocb*) TASK_VAL(task); |
acb = (struct aiocb*) TASK_VAL(task); |
if (acb != ((struct aiocb*) res[i].udata)) |
if (acb != ((struct aiocb*) res[i].udata)) |
continue; |
continue; |
else | else { |
flg++; |
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = (u_long) res[i].fflags; |
|
} |
/* remove user handle */ |
/* remove user handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskAIO]); |
pthread_mutex_lock(&r->root_mtx[taskAIO]); |
Line 855 sched_hook_fetch(void *root, void *arg __unused)
|
Line 1055 sched_hook_fetch(void *root, void *arg __unused)
|
acbs = (struct aiocb**) TASK_VAL(task); |
acbs = (struct aiocb**) TASK_VAL(task); |
if (acbs != ((struct aiocb**) res[i].udata)) |
if (acbs != ((struct aiocb**) res[i].udata)) |
continue; |
continue; |
else | else { |
flg++; |
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = (u_long) res[i].fflags; |
|
} |
/* remove user handle */ |
/* remove user handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskLIO]); |
pthread_mutex_lock(&r->root_mtx[taskLIO]); |
Line 904 sched_hook_fetch(void *root, void *arg __unused)
|
Line 1107 sched_hook_fetch(void *root, void *arg __unused)
|
continue; |
continue; |
else { |
else { |
flg++; |
flg++; |
TASK_DATA(task) = (void*) res[i].data; | TASK_RET(task) = res[i].data; |
TASK_DATLEN(task) = res[i].fflags; | TASK_FLAG(task) = (u_long) res[i].fflags; |
} |
} |
/* remove user handle */ |
/* remove user handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
Line 938 sched_hook_fetch(void *root, void *arg __unused)
|
Line 1141 sched_hook_fetch(void *root, void *arg __unused)
|
LOGERR; |
LOGERR; |
} |
} |
} |
} |
|
#else /* end of kevent dispatcher */ |
|
for (i = 0; i < r->root_kq; i++) { |
|
if (FD_ISSET(i, &rfd) || FD_ISSET(i, &xfd)) { |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { |
|
if (TASK_FD(task) != i) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_FLAG(task) = ioctl(TASK_FD(task), |
|
FIONREAD, &TASK_RET(task)); |
|
} |
|
/* remove read handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREAD]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_read, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREAD]); |
|
#endif |
|
if (r->root_hooks.hook_exec.exception) { |
|
if (r->root_hooks.hook_exec.exception(r, NULL)) { |
|
task->task_type = taskUNUSE; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskUNUSE]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); |
|
#endif |
|
} else { |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
} else { |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
} |
|
/* if match equal to 1, remove resouce */ |
|
if (flg == 1) |
|
FD_CLR(i, &r->root_fds[0]); |
|
} |
|
|
|
if (FD_ISSET(i, &wfd)) { |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { |
|
if (TASK_FD(task) != i) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_FLAG(task) = ioctl(TASK_FD(task), |
|
FIONWRITE, &TASK_RET(task)); |
|
} |
|
/* remove write handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskWRITE]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_write, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskWRITE]); |
|
#endif |
|
if (r->root_hooks.hook_exec.exception) { |
|
if (r->root_hooks.hook_exec.exception(r, NULL)) { |
|
task->task_type = taskUNUSE; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskUNUSE]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); |
|
#endif |
|
} else { |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
} else { |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
} |
|
/* if match equal to 1, remove resouce */ |
|
if (flg == 1) |
|
FD_CLR(i, &r->root_fds[1]); |
|
} |
|
} |
|
|
|
/* optimize select */ |
|
for (i = r->root_kq - 1; i > 2; i--) |
|
if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) |
|
break; |
|
if (i > 2) |
|
r->root_kq = i + 1; |
|
#endif /* KQ_SUPPORT */ |
|
|
|
skip_event: |
/* timer update & put in ready queue */ |
/* timer update & put in ready queue */ |
clock_gettime(CLOCK_MONOTONIC, &now); |
clock_gettime(CLOCK_MONOTONIC, &now); |
|
|
Line 1058 sched_hook_condition(void *root, void *arg)
|
Line 1380 sched_hook_condition(void *root, void *arg)
|
|
|
return (void*) (r->root_cond - *(intptr_t*) arg); |
return (void*) (r->root_cond - *(intptr_t*) arg); |
} |
} |
|
|
|
/* |
|
* sched_hook_rtc() - Default RTC hook |
|
* |
|
* @task = current task |
|
* @arg = unused |
|
* return: <0 errors and 0 ok |
|
*/ |
|
#if defined(HAVE_TIMER_CREATE) && defined(HAVE_TIMER_SETTIME) && defined(HAVE_TIMER_DELETE) |
|
void * |
|
sched_hook_rtc(void *task, void *arg __unused) |
|
{ |
|
sched_task_t *sigt = NULL, *t = task; |
|
struct itimerspec its; |
|
struct sigevent evt; |
|
timer_t tmr; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
memset(&evt, 0, sizeof evt); |
|
evt.sigev_notify = SIGEV_SIGNAL; |
|
evt.sigev_signo = (intptr_t) TASK_DATA(t) + SIGRTMIN; |
|
evt.sigev_value.sival_ptr = TASK_DATA(t); |
|
|
|
if (timer_create(CLOCK_MONOTONIC, &evt, &tmr) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} else |
|
TASK_FLAG(t) = (u_long) tmr; |
|
|
|
if (!(sigt = schedSignal(TASK_ROOT(t), _sched_rtcWrapper, TASK_ARG(t), evt.sigev_signo, |
|
t, (size_t) tmr))) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
timer_delete(tmr); |
|
return (void*) -1; |
|
} else |
|
TASK_RET(t) = (uintptr_t) sigt; |
|
|
|
memset(&its, 0, sizeof its); |
|
its.it_value.tv_sec = t->task_val.ts.tv_sec; |
|
its.it_value.tv_nsec = t->task_val.ts.tv_nsec; |
|
|
|
if (timer_settime(tmr, TIMER_RELTIME, &its, NULL) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
schedCancel(sigt); |
|
timer_delete(tmr); |
|
return (void*) -1; |
|
} |
|
|
|
return NULL; |
|
} |
|
#endif /* HAVE_TIMER_CREATE */ |