|
|
| version 1.5.2.2, 2012/05/03 15:05:09 | version 1.14.2.6, 2012/08/23 02:09:54 |
|---|---|
| Line 107 sched_hook_cancel(void *task, void *arg __unused) | Line 107 sched_hook_cancel(void *task, void *arg __unused) |
| sched_task_t *t = task; | sched_task_t *t = task; |
| struct kevent chg[1]; | struct kevent chg[1]; |
| struct timespec timeout = { 0, 0 }; | struct timespec timeout = { 0, 0 }; |
| #ifdef AIO_SUPPORT | |
| struct aiocb *acb; | |
| #ifdef EVFILT_LIO | |
| register int i = 0; | |
| struct aiocb **acbs; | |
| #endif /* EVFILT_LIO */ | |
| #endif /* AIO_SUPPORT */ | |
| if (!t || !TASK_ROOT(t)) | if (!t || !TASK_ROOT(t)) |
| return (void*) -1; | return (void*) -1; |
| Line 118 sched_hook_cancel(void *task, void *arg __unused) | Line 125 sched_hook_cancel(void *task, void *arg __unused) |
| #else | #else |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
| #endif | #endif |
| kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); | |
| break; | break; |
| case taskWRITE: | case taskWRITE: |
| #ifdef __NetBSD__ | #ifdef __NetBSD__ |
| Line 126 sched_hook_cancel(void *task, void *arg __unused) | Line 132 sched_hook_cancel(void *task, void *arg __unused) |
| #else | #else |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
| #endif | #endif |
| kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); | |
| break; | break; |
| default: | case taskALARM: |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, | |
| 0, 0, (intptr_t) TASK_DATA(t)); | |
| #else | |
| EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, | |
| 0, 0, (void*) TASK_DATA(t)); | |
| #endif | |
| break; | break; |
| case taskNODE: | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); | |
| #endif | |
| break; | |
| case taskPROC: | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| break; | |
| case taskSIGNAL: | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| break; | |
| #ifdef AIO_SUPPORT | |
| case taskAIO: | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| acb = (struct aiocb*) TASK_VAL(t); | |
| if (acb) { | |
| if (aio_cancel(acb->aio_fildes, acb) == AIO_CANCELED) | |
| aio_return(acb); | |
| free(acb); | |
| TASK_VAL(t) = 0; | |
| } | |
| break; | |
| #ifdef EVFILT_LIO | |
| case taskLIO: | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| acbs = (struct aiocb**) TASK_VAL(t); | |
| if (acbs) { | |
| for (i = 0; i < TASK_DATLEN(t); i++) { | |
| if (aio_cancel(acbs[i]->aio_fildes, acbs[i]) == AIO_CANCELED) | |
| aio_return(acbs[i]); | |
| free(acbs[i]); | |
| } | |
| free(acbs); | |
| TASK_VAL(t) = 0; | |
| } | |
| break; | |
| #endif /* EVFILT_LIO */ | |
| #endif /* AIO_SUPPORT */ | |
| #ifdef EVFILT_USER | |
| case taskUSER: | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| break; | |
| #endif | |
| case taskTHREAD: | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_cancel((pthread_t) TASK_VAL(t)); | |
| #endif | |
| default: | |
| return NULL; | |
| } | } |
| kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); | |
| return NULL; | return NULL; |
| } | } |
| #ifdef HAVE_LIBPTHREAD | |
| /* | /* |
| * sched_hook_thread() - Default THREAD hook | |
| * | |
| * @task = current task | |
| * @arg = pthread attributes | |
| * return: <0 errors and 0 ok | |
| */ | |
| void * | |
| sched_hook_thread(void *task, void *arg) | |
| { | |
| sched_task_t *t = task; | |
| pthread_t tid; | |
| sigset_t s, o; | |
| if (!t || !TASK_ROOT(t)) | |
| return (void*) -1; | |
| sigfillset(&s); | |
| pthread_sigmask(SIG_BLOCK, &s, &o); | |
| if (pthread_create(&tid, (pthread_attr_t*) arg, | |
| (void *(*)(void*)) _sched_threadWrapper, t)) { | |
| LOGERR; | |
| pthread_sigmask(SIG_SETMASK, &o, NULL); | |
| return (void*) -1; | |
| } else | |
| TASK_VAL(t) = (u_long) tid; | |
| if (!TASK_ISLOCKED(t)) | |
| TASK_LOCK(t); | |
| pthread_sigmask(SIG_SETMASK, &o, NULL); | |
| return NULL; | |
| } | |
| #endif | |
| /* | |
| * sched_hook_read() - Default READ hook | * sched_hook_read() - Default READ hook |
| * | * |
| * @task = current task | * @task = current task |
| Line 153 sched_hook_read(void *task, void *arg __unused) | Line 273 sched_hook_read(void *task, void *arg __unused) |
| return (void*) -1; | return (void*) -1; |
| #ifdef __NetBSD__ | #ifdef __NetBSD__ |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (intptr_t) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
| #else | #else |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (void*) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_FD(t)); |
| #endif | #endif |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
| Line 186 sched_hook_write(void *task, void *arg __unused) | Line 306 sched_hook_write(void *task, void *arg __unused) |
| return (void*) -1; | return (void*) -1; |
| #ifdef __NetBSD__ | #ifdef __NetBSD__ |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (intptr_t) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); |
| #else | #else |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (void*) TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_FD(t)); |
| #endif | #endif |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | if (TASK_ROOT(t)->root_hooks.hook_exec.exception) |
| Line 202 sched_hook_write(void *task, void *arg __unused) | Line 322 sched_hook_write(void *task, void *arg __unused) |
| } | } |
| /* | /* |
| * sched_hook_alarm() - Default ALARM hook | |
| * | |
| * @task = current task | |
| * @arg = unused | |
| * return: <0 errors and 0 ok | |
| */ | |
| void * | |
| sched_hook_alarm(void *task, void *arg __unused) | |
| { | |
| sched_task_t *t = task; | |
| struct kevent chg[1]; | |
| struct timespec timeout = { 0, 0 }; | |
| if (!t || !TASK_ROOT(t)) | |
| return (void*) -1; | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, | |
| t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, | |
| (intptr_t) TASK_DATA(t)); | |
| #else | |
| EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, | |
| t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, | |
| (void*) TASK_DATA(t)); | |
| #endif | |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | |
| TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); | |
| else | |
| LOGERR; | |
| return (void*) -1; | |
| } | |
| return NULL; | |
| } | |
| /* | |
| * sched_hook_node() - Default NODE hook | |
| * | |
| * @task = current task | |
| * @arg = unused | |
| * return: <0 errors and 0 ok | |
| */ | |
| void * | |
| sched_hook_node(void *task, void *arg __unused) | |
| { | |
| sched_task_t *t = task; | |
| struct kevent chg[1]; | |
| struct timespec timeout = { 0, 0 }; | |
| if (!t || !TASK_ROOT(t)) | |
| return (void*) -1; | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_ADD | EV_CLEAR, | |
| NOTE_DELETE | NOTE_WRITE | NOTE_EXTEND | NOTE_ATTRIB | | |
| NOTE_LINK | NOTE_RENAME | NOTE_REVOKE, 0, (intptr_t) TASK_FD(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_ADD | EV_CLEAR, | |
| NOTE_DELETE | NOTE_WRITE | NOTE_EXTEND | NOTE_ATTRIB | | |
| NOTE_LINK | NOTE_RENAME | NOTE_REVOKE, 0, (void*) TASK_FD(t)); | |
| #endif | |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | |
| TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); | |
| else | |
| LOGERR; | |
| return (void*) -1; | |
| } | |
| return NULL; | |
| } | |
| /* | |
| * sched_hook_proc() - Default PROC hook | |
| * | |
| * @task = current task | |
| * @arg = unused | |
| * return: <0 errors and 0 ok | |
| */ | |
| void * | |
| sched_hook_proc(void *task, void *arg __unused) | |
| { | |
| sched_task_t *t = task; | |
| struct kevent chg[1]; | |
| struct timespec timeout = { 0, 0 }; | |
| if (!t || !TASK_ROOT(t)) | |
| return (void*) -1; | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_ADD | EV_CLEAR, | |
| NOTE_EXIT | NOTE_FORK | NOTE_EXEC | NOTE_TRACK, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_ADD | EV_CLEAR, | |
| NOTE_EXIT | NOTE_FORK | NOTE_EXEC | NOTE_TRACK, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | |
| TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); | |
| else | |
| LOGERR; | |
| return (void*) -1; | |
| } | |
| return NULL; | |
| } | |
| /* | |
| * sched_hook_signal() - Default SIGNAL hook | |
| * | |
| * @task = current task | |
| * @arg = unused | |
| * return: <0 errors and 0 ok | |
| */ | |
| void * | |
| sched_hook_signal(void *task, void *arg __unused) | |
| { | |
| sched_task_t *t = task; | |
| struct kevent chg[1]; | |
| struct timespec timeout = { 0, 0 }; | |
| if (!t || !TASK_ROOT(t)) | |
| return (void*) -1; | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (void*) TASK_VAL(t)); | |
| #endif | |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | |
| TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); | |
| else | |
| LOGERR; | |
| return (void*) -1; | |
| } | |
| return NULL; | |
| } | |
| /* | |
| * sched_hook_user() - Default USER hook | |
| * | |
| * @task = current task | |
| * @arg = unused | |
| * return: <0 errors and 0 ok | |
| */ | |
| #ifdef EVFILT_USER | |
| void * | |
| sched_hook_user(void *task, void *arg __unused) | |
| { | |
| sched_task_t *t = task; | |
| struct kevent chg[1]; | |
| struct timespec timeout = { 0, 0 }; | |
| if (!t || !TASK_ROOT(t)) | |
| return (void*) -1; | |
| #ifdef __NetBSD__ | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_ADD | EV_CLEAR, TASK_DATLEN(t), | |
| 0, (intptr_t) TASK_VAL(t)); | |
| #else | |
| EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_ADD | EV_CLEAR, TASK_DATLEN(t), | |
| 0, (void*) TASK_VAL(t)); | |
| #endif | |
| if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { | |
| if (TASK_ROOT(t)->root_hooks.hook_exec.exception) | |
| TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); | |
| else | |
| LOGERR; | |
| return (void*) -1; | |
| } | |
| return NULL; | |
| } | |
| #endif | |
| /* | |
| * sched_hook_fetch() - Default FETCH hook | * sched_hook_fetch() - Default FETCH hook |
| * | * |
| * @root = root task | * @root = root task |
| Line 212 void * | Line 511 void * |
| sched_hook_fetch(void *root, void *arg __unused) | sched_hook_fetch(void *root, void *arg __unused) |
| { | { |
| sched_root_task_t *r = root; | sched_root_task_t *r = root; |
| sched_task_t *task; | sched_task_t *task, *tmp; |
| struct timespec now, m, mtmp; | struct timespec now, m, mtmp; |
| struct timespec *timeout; | struct timespec *timeout; |
| struct kevent evt[1], res[KQ_EVENTS]; | struct kevent evt[1], res[KQ_EVENTS]; |
| register int i; | register int i, flg; |
| int en; | int en; |
| #ifdef AIO_SUPPORT | |
| int len, fd; | |
| struct aiocb *acb; | |
| #ifdef EVFILT_LIO | |
| int l; | |
| register int j; | |
| off_t off; | |
| struct aiocb **acbs; | |
| struct iovec *iv; | |
| #endif /* EVFILT_LIO */ | |
| #endif /* AIO_SUPPORT */ | |
| if (!r) | if (!r) |
| return NULL; | return NULL; |
| /* get new task by queue priority */ | /* get new task by queue priority */ |
| retry: | |
| while ((task = TAILQ_FIRST(&r->root_event))) { | while ((task = TAILQ_FIRST(&r->root_event))) { |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| pthread_mutex_lock(&r->root_mtx[taskEVENT]); | pthread_mutex_lock(&r->root_mtx[taskEVENT]); |
| Line 281 retry: | Line 590 retry: |
| sched_timespecinf(&r->root_wait); | sched_timespecinf(&r->root_wait); |
| } | } |
| #else | #else |
| if (!TAILQ_FIRST(&r->root_eventlo) && (task = TAILQ_FIRST(&r->root_timer))) { | if (!TAILQ_FIRST(&r->root_task) && (task = TAILQ_FIRST(&r->root_timer))) { |
| clock_gettime(CLOCK_MONOTONIC, &now); | clock_gettime(CLOCK_MONOTONIC, &now); |
| m = TASK_TS(task); | m = TASK_TS(task); |
| Line 292 retry: | Line 601 retry: |
| sched_timespecinf(&r->root_wait); | sched_timespecinf(&r->root_wait); |
| } | } |
| #endif | #endif |
| /* if present member of eventLo, set NOWAIT */ | /* if present member of task, set NOWAIT */ |
| if (TAILQ_FIRST(&r->root_eventlo)) | if (TAILQ_FIRST(&r->root_task)) |
| sched_timespecclear(&r->root_wait); | sched_timespecclear(&r->root_wait); |
| if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) | if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) |
| Line 306 retry: | Line 615 retry: |
| if (r->root_hooks.hook_exec.exception) { | if (r->root_hooks.hook_exec.exception) { |
| if (r->root_hooks.hook_exec.exception(r, NULL)) | if (r->root_hooks.hook_exec.exception(r, NULL)) |
| return NULL; | return NULL; |
| } else | } else if (errno != EINTR) |
| LOGERR; | LOGERR; |
| #ifdef NDEBUG | |
| /* kevent no exit by error, if non-debug version */ | |
| goto retry; | |
| #else | |
| /* diagnostic exit from scheduler if kevent error occur */ | |
| return NULL; | return NULL; |
| #endif | |
| } | } |
| now.tv_sec = now.tv_nsec = 0; | now.tv_sec = now.tv_nsec = 0; |
| Line 325 retry: | Line 628 retry: |
| /* Put read/write task to ready queue */ | /* Put read/write task to ready queue */ |
| switch (res[i].filter) { | switch (res[i].filter) { |
| case EVFILT_READ: | case EVFILT_READ: |
| TAILQ_FOREACH(task, &r->root_read, task_node) { | flg = 0; |
| TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { | |
| if (TASK_FD(task) != ((intptr_t) res[i].udata)) | if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
| continue; | continue; |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove read handle */ | /* remove read handle */ |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| pthread_mutex_lock(&r->root_mtx[taskREAD]); | pthread_mutex_lock(&r->root_mtx[taskREAD]); |
| Line 366 retry: | Line 675 retry: |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
| #endif | #endif |
| } | } |
| break; | |
| } | } |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | break; |
| case EVFILT_WRITE: | case EVFILT_WRITE: |
| TAILQ_FOREACH(task, &r->root_write, task_node) { | flg = 0; |
| TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { | |
| if (TASK_FD(task) != ((intptr_t) res[i].udata)) | if (TASK_FD(task) != ((intptr_t) res[i].udata)) |
| continue; | continue; |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove write handle */ | /* remove write handle */ |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| pthread_mutex_lock(&r->root_mtx[taskWRITE]); | pthread_mutex_lock(&r->root_mtx[taskWRITE]); |
| Line 411 retry: | Line 728 retry: |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
| #endif | #endif |
| } | } |
| break; | |
| } | } |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | break; |
| case EVFILT_TIMER: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_alarm, task_node, tmp) { | |
| if ((uintptr_t) TASK_DATA(task) != ((uintptr_t) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove alarm handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskALARM]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_alarm, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskALARM]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| case EVFILT_VNODE: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_node, task_node, tmp) { | |
| if (TASK_FD(task) != ((intptr_t) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove node handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskNODE]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_node, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskNODE]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| case EVFILT_PROC: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_proc, task_node, tmp) { | |
| if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove proc handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskPROC]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_proc, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskPROC]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| case EVFILT_SIGNAL: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_signal, task_node, tmp) { | |
| if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove signal handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskSIGNAL]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_signal, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskSIGNAL]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| #ifdef AIO_SUPPORT | |
| case EVFILT_AIO: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_aio, task_node, tmp) { | |
| acb = (struct aiocb*) TASK_VAL(task); | |
| if (acb != ((struct aiocb*) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove user handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskAIO]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_aio, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskAIO]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| fd = acb->aio_fildes; | |
| if ((len = aio_return(acb)) != -1) { | |
| if (lseek(fd, acb->aio_offset + len, SEEK_CUR) == -1) | |
| LOGERR; | |
| } else | |
| LOGERR; | |
| free(acb); | |
| TASK_DATLEN(task) = (u_long) len; | |
| TASK_FD(task) = fd; | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| #ifdef EVFILT_LIO | |
| case EVFILT_LIO: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_lio, task_node, tmp) { | |
| acbs = (struct aiocb**) TASK_VAL(task); | |
| if (acbs != ((struct aiocb**) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove user handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskLIO]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_lio, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskLIO]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| iv = (struct iovec*) TASK_DATA(task); | |
| fd = acbs[0]->aio_fildes; | |
| off = acbs[0]->aio_offset; | |
| for (j = len = 0; i < TASK_DATLEN(task); len += l, i++) { | |
| if ((iv[i].iov_len = aio_return(acbs[i])) == -1) | |
| l = 0; | |
| else | |
| l = iv[i].iov_len; | |
| free(acbs[i]); | |
| } | |
| free(acbs); | |
| TASK_DATLEN(task) = (u_long) len; | |
| TASK_FD(task) = fd; | |
| if (lseek(fd, off + len, SEEK_CUR) == -1) | |
| LOGERR; | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| #endif /* EVFILT_LIO */ | |
| #endif /* AIO_SUPPORT */ | |
| #ifdef EVFILT_USER | |
| case EVFILT_USER: | |
| flg = 0; | |
| TAILQ_FOREACH_SAFE(task, &r->root_user, task_node, tmp) { | |
| if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) | |
| continue; | |
| else { | |
| flg++; | |
| TASK_RET(task) = res[i].data; | |
| TASK_FLAG(task) = res[i].fflags; | |
| } | |
| /* remove user handle */ | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskUSER]); | |
| #endif | |
| TAILQ_REMOVE(&r->root_user, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskUSER]); | |
| #endif | |
| task->task_type = taskREADY; | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_lock(&r->root_mtx[taskREADY]); | |
| #endif | |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | |
| #ifdef HAVE_LIBPTHREAD | |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | |
| #endif | |
| } | |
| /* if match at least 2, don't remove resouce of event */ | |
| if (flg > 1) | |
| evt->flags ^= evt->flags; | |
| break; | |
| #endif /* EVFILT_USER */ | |
| } | } |
| if (kevent(r->root_kq, evt, 1, NULL, 0, &now) == -1) { | if (kevent(r->root_kq, evt, 1, NULL, 0, &now) == -1) { |
| if (r->root_hooks.hook_exec.exception) { | if (r->root_hooks.hook_exec.exception) { |
| Line 427 retry: | Line 996 retry: |
| /* timer update & put in ready queue */ | /* timer update & put in ready queue */ |
| clock_gettime(CLOCK_MONOTONIC, &now); | clock_gettime(CLOCK_MONOTONIC, &now); |
| TAILQ_FOREACH(task, &r->root_timer, task_node) | TAILQ_FOREACH_SAFE(task, &r->root_timer, task_node, tmp) |
| if (sched_timespeccmp(&now, &TASK_TS(task), -) >= 0) { | if (sched_timespeccmp(&now, &TASK_TS(task), -) >= 0) { |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| pthread_mutex_lock(&r->root_mtx[taskTIMER]); | pthread_mutex_lock(&r->root_mtx[taskTIMER]); |
| Line 446 retry: | Line 1015 retry: |
| #endif | #endif |
| } | } |
| /* put eventlo priority task to ready queue, if there is no ready task or | /* put regular task priority task to ready queue, |
| reach max missed fetch-rotate */ | if there is no ready task or reach max missing hit for regular task */ |
| if ((task = TAILQ_FIRST(&r->root_eventlo))) { | if ((task = TAILQ_FIRST(&r->root_task))) { |
| if (!TAILQ_FIRST(&r->root_ready) || r->root_eventlo_miss > MAX_EVENTLO_MISS) { | if (!TAILQ_FIRST(&r->root_ready) || r->root_miss >= TASK_VAL(task)) { |
| r->root_eventlo_miss = 0; | r->root_miss ^= r->root_miss; |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| pthread_mutex_lock(&r->root_mtx[taskEVENTLO]); | pthread_mutex_lock(&r->root_mtx[taskTASK]); |
| #endif | #endif |
| TAILQ_REMOVE(&r->root_eventlo, task, task_node); | TAILQ_REMOVE(&r->root_task, task, task_node); |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| pthread_mutex_unlock(&r->root_mtx[taskEVENTLO]); | pthread_mutex_unlock(&r->root_mtx[taskTASK]); |
| #endif | #endif |
| task->task_type = taskREADY; | task->task_type = taskREADY; |
| #ifdef HAVE_LIBPTHREAD | #ifdef HAVE_LIBPTHREAD |
| Line 468 retry: | Line 1037 retry: |
| pthread_mutex_unlock(&r->root_mtx[taskREADY]); | pthread_mutex_unlock(&r->root_mtx[taskREADY]); |
| #endif | #endif |
| } else | } else |
| r->root_eventlo_miss++; | r->root_miss++; |
| } else | } else |
| r->root_eventlo_miss = 0; | r->root_miss ^= r->root_miss; |
| /* OK, lets get ready task !!! */ | /* OK, lets get ready task !!! */ |
| task = TAILQ_FIRST(&r->root_ready); | task = TAILQ_FIRST(&r->root_ready); |