version 1.5.2.2, 2012/05/03 15:05:09
|
version 1.13.2.2, 2012/08/21 11:45:35
|
Line 107 sched_hook_cancel(void *task, void *arg __unused)
|
Line 107 sched_hook_cancel(void *task, void *arg __unused)
|
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout = { 0, 0 }; |
struct timespec timeout = { 0, 0 }; |
|
#ifdef AIO_SUPPORT |
|
struct aiocb *acb; |
|
#ifdef EVFILT_LIO |
|
struct aiocb **acbs; |
|
register int i; |
|
#endif /* EVFILT_LIO */ |
|
#endif /* AIO_SUPPORT */ |
|
|
if (!t || !TASK_ROOT(t)) |
if (!t || !TASK_ROOT(t)) |
return (void*) -1; |
return (void*) -1; |
Line 118 sched_hook_cancel(void *task, void *arg __unused)
|
Line 125 sched_hook_cancel(void *task, void *arg __unused)
|
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); |
|
break; |
break; |
case taskWRITE: |
case taskWRITE: |
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
Line 126 sched_hook_cancel(void *task, void *arg __unused)
|
Line 132 sched_hook_cancel(void *task, void *arg __unused)
|
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); |
|
break; |
break; |
default: | case taskALARM: |
| #ifdef __NetBSD__ |
| EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, |
| 0, 0, (intptr_t) TASK_DATA(t)); |
| #else |
| EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, |
| 0, 0, (void*) TASK_DATA(t)); |
| #endif |
break; |
break; |
|
case taskNODE: |
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
|
#endif |
|
break; |
|
case taskPROC: |
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
break; |
|
case taskSIGNAL: |
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
break; |
|
#ifdef AIO_SUPPORT |
|
case taskAIO: |
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
acb = (struct aiocb*) TASK_VAL(t); |
|
if (acb) { |
|
if (aio_cancel(acb->aio_fildes, acb) == AIO_CANCELED) |
|
aio_return(acb); |
|
free(acb); |
|
TASK_VAL(t) = 0; |
|
} |
|
break; |
|
#ifdef EVFILT_LIO |
|
case taskLIO: |
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
acbs = (struct aiocb**) TASK_VAL(t); |
|
if (acbs) { |
|
for (i = 0; i < TASK_DATLEN(t); i++) { |
|
if (aio_cancel(acbs[i]->aio_fildes, acbs[i]) == AIO_CANCELED) |
|
aio_return(acbs[i]); |
|
free(acbs[i]); |
|
} |
|
free(acbs); |
|
TASK_VAL(t) = 0; |
|
} |
|
break; |
|
#endif /* EVFILT_LIO */ |
|
#endif /* AIO_SUPPORT */ |
|
#ifdef EVFILT_USER |
|
case taskUSER: |
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
break; |
|
#endif |
|
case taskTHREAD: |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_cancel((pthread_t) TASK_VAL(t)); |
|
#endif |
|
default: |
|
return NULL; |
} |
} |
|
|
|
kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); |
return NULL; |
return NULL; |
} |
} |
|
|
Line 153 sched_hook_read(void *task, void *arg __unused)
|
Line 237 sched_hook_read(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
|
|
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (intptr_t) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (void*) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
Line 186 sched_hook_write(void *task, void *arg __unused)
|
Line 270 sched_hook_write(void *task, void *arg __unused)
|
return (void*) -1; |
return (void*) -1; |
|
|
#ifdef __NetBSD__ |
#ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (intptr_t) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
#else |
#else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (void*) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_FD(t)); |
#endif |
#endif |
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
Line 202 sched_hook_write(void *task, void *arg __unused)
|
Line 286 sched_hook_write(void *task, void *arg __unused)
|
} |
} |
|
|
/* |
/* |
|
* sched_hook_alarm() - Default ALARM hook |
|
* |
|
* @task = current task |
|
* @arg = unused |
|
* return: <0 errors and 0 ok |
|
*/ |
|
void * |
|
sched_hook_alarm(void *task, void *arg __unused) |
|
{ |
|
sched_task_t *t = task; |
|
struct kevent chg[1]; |
|
struct timespec timeout = { 0, 0 }; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, |
|
t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, |
|
(intptr_t) TASK_DATA(t)); |
|
#else |
|
EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, |
|
t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, |
|
(void*) TASK_DATA(t)); |
|
#endif |
|
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} |
|
|
|
return NULL; |
|
} |
|
|
|
/* |
|
* sched_hook_node() - Default NODE hook |
|
* |
|
* @task = current task |
|
* @arg = unused |
|
* return: <0 errors and 0 ok |
|
*/ |
|
void * |
|
sched_hook_node(void *task, void *arg __unused) |
|
{ |
|
sched_task_t *t = task; |
|
struct kevent chg[1]; |
|
struct timespec timeout = { 0, 0 }; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_ADD | EV_CLEAR, |
|
NOTE_DELETE | NOTE_WRITE | NOTE_EXTEND | NOTE_ATTRIB | |
|
NOTE_LINK | NOTE_RENAME | NOTE_REVOKE, 0, (intptr_t) TASK_FD(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_ADD | EV_CLEAR, |
|
NOTE_DELETE | NOTE_WRITE | NOTE_EXTEND | NOTE_ATTRIB | |
|
NOTE_LINK | NOTE_RENAME | NOTE_REVOKE, 0, (void*) TASK_FD(t)); |
|
#endif |
|
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} |
|
|
|
return NULL; |
|
} |
|
|
|
/* |
|
* sched_hook_proc() - Default PROC hook |
|
* |
|
* @task = current task |
|
* @arg = unused |
|
* return: <0 errors and 0 ok |
|
*/ |
|
void * |
|
sched_hook_proc(void *task, void *arg __unused) |
|
{ |
|
sched_task_t *t = task; |
|
struct kevent chg[1]; |
|
struct timespec timeout = { 0, 0 }; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_ADD | EV_CLEAR, |
|
NOTE_EXIT | NOTE_FORK | NOTE_EXEC | NOTE_TRACK, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_ADD | EV_CLEAR, |
|
NOTE_EXIT | NOTE_FORK | NOTE_EXEC | NOTE_TRACK, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} |
|
|
|
return NULL; |
|
} |
|
|
|
/* |
|
* sched_hook_signal() - Default SIGNAL hook |
|
* |
|
* @task = current task |
|
* @arg = unused |
|
* return: <0 errors and 0 ok |
|
*/ |
|
void * |
|
sched_hook_signal(void *task, void *arg __unused) |
|
{ |
|
sched_task_t *t = task; |
|
struct kevent chg[1]; |
|
struct timespec timeout = { 0, 0 }; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (void*) TASK_VAL(t)); |
|
#endif |
|
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} |
|
|
|
return NULL; |
|
} |
|
|
|
/* |
|
* sched_hook_user() - Default USER hook |
|
* |
|
* @task = current task |
|
* @arg = unused |
|
* return: <0 errors and 0 ok |
|
*/ |
|
#ifdef EVFILT_USER |
|
void * |
|
sched_hook_user(void *task, void *arg __unused) |
|
{ |
|
sched_task_t *t = task; |
|
struct kevent chg[1]; |
|
struct timespec timeout = { 0, 0 }; |
|
|
|
if (!t || !TASK_ROOT(t)) |
|
return (void*) -1; |
|
|
|
#ifdef __NetBSD__ |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_ADD | EV_CLEAR, TASK_DATLEN(t), |
|
0, (intptr_t) TASK_VAL(t)); |
|
#else |
|
EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_ADD | EV_CLEAR, TASK_DATLEN(t), |
|
0, (void*) TASK_VAL(t)); |
|
#endif |
|
if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
|
if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
|
TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); |
|
else |
|
LOGERR; |
|
return (void*) -1; |
|
} |
|
|
|
return NULL; |
|
} |
|
#endif |
|
|
|
/* |
* sched_hook_fetch() - Default FETCH hook |
* sched_hook_fetch() - Default FETCH hook |
* |
* |
* @root = root task |
* @root = root task |
Line 212 void *
|
Line 475 void *
|
sched_hook_fetch(void *root, void *arg __unused) |
sched_hook_fetch(void *root, void *arg __unused) |
{ |
{ |
sched_root_task_t *r = root; |
sched_root_task_t *r = root; |
sched_task_t *task; | sched_task_t *task, *tmp; |
struct timespec now, m, mtmp; |
struct timespec now, m, mtmp; |
struct timespec *timeout; |
struct timespec *timeout; |
struct kevent evt[1], res[KQ_EVENTS]; |
struct kevent evt[1], res[KQ_EVENTS]; |
register int i; | register int i, flg; |
int en; |
int en; |
|
#ifdef AIO_SUPPORT |
|
int len, fd; |
|
struct aiocb *acb; |
|
#ifdef EVFILT_LIO |
|
int l; |
|
register int j; |
|
off_t off; |
|
struct aiocb **acbs; |
|
struct iovec *iv; |
|
#endif /* EVFILT_LIO */ |
|
#endif /* AIO_SUPPORT */ |
|
|
if (!r) |
if (!r) |
return NULL; |
return NULL; |
|
|
/* get new task by queue priority */ |
/* get new task by queue priority */ |
retry: |
|
while ((task = TAILQ_FIRST(&r->root_event))) { |
while ((task = TAILQ_FIRST(&r->root_event))) { |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskEVENT]); |
pthread_mutex_lock(&r->root_mtx[taskEVENT]); |
Line 281 retry:
|
Line 554 retry:
|
sched_timespecinf(&r->root_wait); |
sched_timespecinf(&r->root_wait); |
} |
} |
#else |
#else |
if (!TAILQ_FIRST(&r->root_eventlo) && (task = TAILQ_FIRST(&r->root_timer))) { | if (!TAILQ_FIRST(&r->root_task) && (task = TAILQ_FIRST(&r->root_timer))) { |
clock_gettime(CLOCK_MONOTONIC, &now); |
clock_gettime(CLOCK_MONOTONIC, &now); |
|
|
m = TASK_TS(task); |
m = TASK_TS(task); |
Line 292 retry:
|
Line 565 retry:
|
sched_timespecinf(&r->root_wait); |
sched_timespecinf(&r->root_wait); |
} |
} |
#endif |
#endif |
/* if present member of eventLo, set NOWAIT */ | /* if present member of task, set NOWAIT */ |
if (TAILQ_FIRST(&r->root_eventlo)) | if (TAILQ_FIRST(&r->root_task)) |
sched_timespecclear(&r->root_wait); |
sched_timespecclear(&r->root_wait); |
|
|
if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) |
if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) |
Line 306 retry:
|
Line 579 retry:
|
if (r->root_hooks.hook_exec.exception) { |
if (r->root_hooks.hook_exec.exception) { |
if (r->root_hooks.hook_exec.exception(r, NULL)) |
if (r->root_hooks.hook_exec.exception(r, NULL)) |
return NULL; |
return NULL; |
} else | } else if (errno != EINTR) |
LOGERR; |
LOGERR; |
#ifdef NDEBUG |
|
/* kevent no exit by error, if non-debug version */ |
|
goto retry; |
|
#else |
|
/* diagnostic exit from scheduler if kevent error occur */ |
|
return NULL; |
return NULL; |
#endif |
|
} |
} |
|
|
now.tv_sec = now.tv_nsec = 0; |
now.tv_sec = now.tv_nsec = 0; |
Line 325 retry:
|
Line 592 retry:
|
/* Put read/write task to ready queue */ |
/* Put read/write task to ready queue */ |
switch (res[i].filter) { |
switch (res[i].filter) { |
case EVFILT_READ: |
case EVFILT_READ: |
TAILQ_FOREACH(task, &r->root_read, task_node) { | flg = 0; |
| TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
continue; |
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
/* remove read handle */ |
/* remove read handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskREAD]); |
pthread_mutex_lock(&r->root_mtx[taskREAD]); |
Line 366 retry:
|
Line 639 retry:
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
#endif |
#endif |
} |
} |
break; |
|
} |
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
break; |
break; |
case EVFILT_WRITE: |
case EVFILT_WRITE: |
TAILQ_FOREACH(task, &r->root_write, task_node) { | flg = 0; |
| TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
continue; |
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
/* remove write handle */ |
/* remove write handle */ |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskWRITE]); |
pthread_mutex_lock(&r->root_mtx[taskWRITE]); |
Line 411 retry:
|
Line 692 retry:
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
#endif |
#endif |
} |
} |
break; |
|
} |
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
break; |
break; |
|
case EVFILT_TIMER: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_alarm, task_node, tmp) { |
|
if ((uintptr_t) TASK_DATA(task) != ((uintptr_t) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove alarm handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskALARM]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_alarm, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskALARM]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
case EVFILT_VNODE: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_node, task_node, tmp) { |
|
if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove node handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskNODE]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_node, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskNODE]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
case EVFILT_PROC: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_proc, task_node, tmp) { |
|
if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove proc handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskPROC]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_proc, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskPROC]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
case EVFILT_SIGNAL: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_signal, task_node, tmp) { |
|
if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove signal handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskSIGNAL]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_signal, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskSIGNAL]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
#ifdef AIO_SUPPORT |
|
case EVFILT_AIO: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_aio, task_node, tmp) { |
|
acb = (struct aiocb*) TASK_VAL(task); |
|
if (acb != ((struct aiocb*) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove user handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskAIO]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_aio, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskAIO]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
fd = acb->aio_fildes; |
|
if ((len = aio_return(acb)) != -1) { |
|
if (lseek(fd, acb->aio_offset + len, SEEK_CUR) == -1) |
|
LOGERR; |
|
} else |
|
LOGERR; |
|
free(acb); |
|
TASK_DATLEN(task) = (u_long) len; |
|
TASK_FD(task) = fd; |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
#ifdef EVFILT_LIO |
|
case EVFILT_LIO: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_lio, task_node, tmp) { |
|
acbs = (struct aiocb**) TASK_VAL(task); |
|
if (acbs != ((struct aiocb**) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove user handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskLIO]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_lio, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskLIO]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
iv = (struct iovec*) TASK_DATA(task); |
|
fd = acbs[0]->aio_fildes; |
|
off = acbs[0]->aio_offset; |
|
for (j = len = 0; i < TASK_DATLEN(task); len += l, i++) { |
|
if ((iv[i].iov_len = aio_return(acbs[i])) == -1) |
|
l = 0; |
|
else |
|
l = iv[i].iov_len; |
|
free(acbs[i]); |
|
} |
|
free(acbs); |
|
TASK_DATLEN(task) = (u_long) len; |
|
TASK_FD(task) = fd; |
|
|
|
if (lseek(fd, off + len, SEEK_CUR) == -1) |
|
LOGERR; |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
#endif /* EVFILT_LIO */ |
|
#endif /* AIO_SUPPORT */ |
|
#ifdef EVFILT_USER |
|
case EVFILT_USER: |
|
flg = 0; |
|
TAILQ_FOREACH_SAFE(task, &r->root_user, task_node, tmp) { |
|
if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) |
|
continue; |
|
else { |
|
flg++; |
|
TASK_RET(task) = res[i].data; |
|
TASK_FLAG(task) = res[i].fflags; |
|
} |
|
/* remove user handle */ |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskUSER]); |
|
#endif |
|
TAILQ_REMOVE(&r->root_user, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskUSER]); |
|
#endif |
|
task->task_type = taskREADY; |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_lock(&r->root_mtx[taskREADY]); |
|
#endif |
|
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
|
#ifdef HAVE_LIBPTHREAD |
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
|
#endif |
|
} |
|
/* if match at least 2, don't remove resouce of event */ |
|
if (flg > 1) |
|
evt->flags ^= evt->flags; |
|
break; |
|
#endif /* EVFILT_USER */ |
} |
} |
if (kevent(r->root_kq, evt, 1, NULL, 0, &now) == -1) { |
if (kevent(r->root_kq, evt, 1, NULL, 0, &now) == -1) { |
if (r->root_hooks.hook_exec.exception) { |
if (r->root_hooks.hook_exec.exception) { |
Line 427 retry:
|
Line 960 retry:
|
/* timer update & put in ready queue */ |
/* timer update & put in ready queue */ |
clock_gettime(CLOCK_MONOTONIC, &now); |
clock_gettime(CLOCK_MONOTONIC, &now); |
|
|
TAILQ_FOREACH(task, &r->root_timer, task_node) | TAILQ_FOREACH_SAFE(task, &r->root_timer, task_node, tmp) |
if (sched_timespeccmp(&now, &TASK_TS(task), -) >= 0) { |
if (sched_timespeccmp(&now, &TASK_TS(task), -) >= 0) { |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskTIMER]); |
pthread_mutex_lock(&r->root_mtx[taskTIMER]); |
Line 446 retry:
|
Line 979 retry:
|
#endif |
#endif |
} |
} |
|
|
/* put eventlo priority task to ready queue, if there is no ready task or | /* put regular task priority task to ready queue, |
reach max missed fetch-rotate */ | if there is no ready task or reach max missing hit for regular task */ |
if ((task = TAILQ_FIRST(&r->root_eventlo))) { | if ((task = TAILQ_FIRST(&r->root_task))) { |
if (!TAILQ_FIRST(&r->root_ready) || r->root_eventlo_miss > MAX_EVENTLO_MISS) { | if (!TAILQ_FIRST(&r->root_ready) || r->root_miss >= TASK_VAL(task)) { |
r->root_eventlo_miss = 0; | r->root_miss ^= r->root_miss; |
|
|
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_lock(&r->root_mtx[taskEVENTLO]); | pthread_mutex_lock(&r->root_mtx[taskTASK]); |
#endif |
#endif |
TAILQ_REMOVE(&r->root_eventlo, task, task_node); | TAILQ_REMOVE(&r->root_task, task, task_node); |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
pthread_mutex_unlock(&r->root_mtx[taskEVENTLO]); | pthread_mutex_unlock(&r->root_mtx[taskTASK]); |
#endif |
#endif |
task->task_type = taskREADY; |
task->task_type = taskREADY; |
#ifdef HAVE_LIBPTHREAD |
#ifdef HAVE_LIBPTHREAD |
Line 468 retry:
|
Line 1001 retry:
|
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
#endif |
#endif |
} else |
} else |
r->root_eventlo_miss++; | r->root_miss++; |
} else |
} else |
r->root_eventlo_miss = 0; | r->root_miss ^= r->root_miss; |
|
|
/* OK, lets get ready task !!! */ |
/* OK, lets get ready task !!! */ |
task = TAILQ_FIRST(&r->root_ready); |
task = TAILQ_FIRST(&r->root_ready); |