--- libaitsched/src/hooks.c 2017/09/04 08:47:43 1.32 +++ libaitsched/src/hooks.c 2022/10/19 01:45:08 1.37 @@ -3,7 +3,7 @@ * by Michael Pounov * * $Author: misho $ -* $Id: hooks.c,v 1.32 2017/09/04 08:47:43 misho Exp $ +* $Id: hooks.c,v 1.37 2022/10/19 01:45:08 misho Exp $ * ************************************************************************** The ELWIX and AITNET software is distributed under the following @@ -12,7 +12,7 @@ terms: All of the documentation and software included in the ELWIX and AITNET Releases is copyrighted by ELWIX - Sofia/Bulgaria -Copyright 2004 - 2017 +Copyright 2004 - 2022 by Michael Pounov . All rights reserved. Redistribution and use in source and binary forms, with or without @@ -204,7 +204,7 @@ sched_hook_cancel(void *task, void *arg __unused) { sched_task_t *t = task, *tmp, *tt; sched_root_task_t *r = NULL; - int flg; + int flg = 0; #if SUP_ENABLE == KQ_SUPPORT struct kevent chg[1]; struct timespec timeout = { 0, 0 }; @@ -229,11 +229,8 @@ sched_hook_cancel(void *task, void *arg __unused) switch (TASK_TYPE(t)) { case taskREAD: /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_read, task_node, tmp) - if (TASK_FD(tt) != TASK_FD(t)) - continue; - else + if (TASK_FD(tt) == TASK_FD(t)) flg++; #if SUP_ENABLE == KQ_SUPPORT #ifdef __NetBSD__ @@ -258,21 +255,17 @@ sched_hook_cancel(void *task, void *arg __unused) FD_CLR(TASK_FD(t), &r->root_fds[0]); /* optimize select */ - for (i = r->root_kq - 1; i > 2; i--) + for (i = r->root_kq - 1; i >= 0; i--) if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) break; - if (i > 2) - r->root_kq = i + 1; + r->root_kq = i + 1; } #endif break; case taskWRITE: /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_write, task_node, tmp) - if (TASK_FD(tt) != TASK_FD(t)) - continue; - else + if (TASK_FD(tt) == TASK_FD(t)) flg++; #if SUP_ENABLE == KQ_SUPPORT #ifdef __NetBSD__ @@ -297,22 +290,18 @@ sched_hook_cancel(void *task, void *arg __unused) FD_CLR(TASK_FD(t), &r->root_fds[1]); /* optimize select */ - for (i = r->root_kq - 1; i > 2; i--) + for (i = r->root_kq - 1; i >= 0; i--) if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) break; - if (i > 2) - r->root_kq = i + 1; + r->root_kq = i + 1; } #endif break; case taskALARM: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_alarm, task_node, tmp) - if (TASK_DATA(tt) != TASK_DATA(t)) - continue; - else + if (TASK_DATA(tt) == TASK_DATA(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, flg < 2 ? EV_DELETE : 0, @@ -326,11 +315,8 @@ sched_hook_cancel(void *task, void *arg __unused) case taskNODE: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_node, task_node, tmp) - if (TASK_FD(tt) != TASK_FD(t)) - continue; - else + if (TASK_FD(tt) == TASK_FD(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, flg < 2 ? EV_DELETE : 0, @@ -344,11 +330,8 @@ sched_hook_cancel(void *task, void *arg __unused) case taskPROC: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_proc, task_node, tmp) - if (TASK_VAL(tt) != TASK_VAL(t)) - continue; - else + if (TASK_VAL(tt) == TASK_VAL(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, flg < 2 ? EV_DELETE : 0, @@ -362,11 +345,8 @@ sched_hook_cancel(void *task, void *arg __unused) case taskSIGNAL: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_signal, task_node, tmp) - if (TASK_VAL(tt) != TASK_VAL(t)) - continue; - else + if (TASK_VAL(tt) == TASK_VAL(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, flg < 2 ? EV_DELETE : 0, @@ -384,11 +364,8 @@ sched_hook_cancel(void *task, void *arg __unused) case taskAIO: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_aio, task_node, tmp) - if (TASK_VAL(tt) != TASK_VAL(t)) - continue; - else + if (TASK_VAL(tt) == TASK_VAL(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, flg < 2 ? EV_DELETE : 0, @@ -401,7 +378,7 @@ sched_hook_cancel(void *task, void *arg __unused) if (acb) { if (aio_cancel(acb->aio_fildes, acb) == AIO_CANCELED) aio_return(acb); - free(acb); + e_free(acb); TASK_VAL(t) = 0; } #endif @@ -410,11 +387,8 @@ sched_hook_cancel(void *task, void *arg __unused) case taskLIO: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_lio, task_node, tmp) - if (TASK_VAL(tt) != TASK_VAL(t)) - continue; - else + if (TASK_VAL(tt) == TASK_VAL(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, flg < 2 ? EV_DELETE : 0, @@ -428,9 +402,9 @@ sched_hook_cancel(void *task, void *arg __unused) for (i = 0; i < TASK_DATLEN(t); i++) { if (aio_cancel(acbs[i]->aio_fildes, acbs[i]) == AIO_CANCELED) aio_return(acbs[i]); - free(acbs[i]); + e_free(acbs[i]); } - free(acbs); + e_free(acbs); TASK_VAL(t) = 0; } #endif @@ -441,11 +415,8 @@ sched_hook_cancel(void *task, void *arg __unused) case taskUSER: #if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_user, task_node, tmp) - if (TASK_VAL(tt) != TASK_VAL(t)) - continue; - else + if (TASK_VAL(tt) == TASK_VAL(t)) flg++; #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, flg < 2 ? EV_DELETE : 0, @@ -477,11 +448,8 @@ sched_hook_cancel(void *task, void *arg __unused) schedCancel((sched_task_t*) TASK_RET(t)); #else /* check for multi subscribers */ - flg = 0; TAILQ_FOREACH_SAFE(tt, &r->root_rtc, task_node, tmp) - if (TASK_DATA(tt) != TASK_DATA(t)) - continue; - else + if (TASK_DATA(tt) == TASK_DATA(t)) flg++; /* restore signal */ @@ -497,7 +465,9 @@ sched_hook_cancel(void *task, void *arg __unused) #if SUP_ENABLE == KQ_SUPPORT kevent(r->root_kq, chg, 1, NULL, 0, &timeout); #elif SUP_ENABLE == EP_SUPPORT - epoll_ctl(r->root_kq, ee.events ? EPOLL_CTL_MOD : EPOLL_CTL_DEL, ee.data.fd, &ee); + if (TASK_TYPE(t) == taskREAD || TASK_TYPE(t) == taskWRITE) { + epoll_ctl(r->root_kq, ee.events ? EPOLL_CTL_MOD : EPOLL_CTL_DEL, ee.data.fd, &ee); + } #endif return NULL; } @@ -650,8 +620,9 @@ sched_hook_write(void *task, void *arg __unused) flg |= 1; ee.events |= EPOLLIN | EPOLLPRI; } - if (FD_ISSET(TASK_FD(t), &r->root_fds[1])) + if (FD_ISSET(TASK_FD(t), &r->root_fds[1])) { flg |= 2; + } if (epoll_ctl(r->root_kq, flg ? EPOLL_CTL_MOD : EPOLL_CTL_ADD, TASK_FD(t), &ee) == -1) { if (r->root_hooks.hook_exec.exception) @@ -713,16 +684,17 @@ sched_hook_alarm(void *task, void *arg __unused) * sched_hook_node() - Default NODE hook * * @task = current task - * @arg = unused + * @arg = if arg == 42 then waiting for all events * return: <0 errors and 0 ok */ void * -sched_hook_node(void *task, void *arg __unused) +sched_hook_node(void *task, void *arg) { #if SUP_ENABLE == KQ_SUPPORT sched_task_t *t = task; struct kevent chg[1]; struct timespec timeout = { 0, 0 }; + u_int addflags = (u_int) (uintptr_t) arg; if (!t || !TASK_ROOT(t)) return (void*) -1; @@ -730,11 +702,11 @@ sched_hook_node(void *task, void *arg __unused) #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_ADD | EV_CLEAR, NOTE_DELETE | NOTE_WRITE | NOTE_EXTEND | NOTE_ATTRIB | - NOTE_LINK | NOTE_RENAME | NOTE_REVOKE, 0, (intptr_t) TASK_FD(t)); + NOTE_LINK | NOTE_RENAME | NOTE_REVOKE | addflags, 0, (intptr_t) TASK_FD(t)); #else EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_ADD | EV_CLEAR, NOTE_DELETE | NOTE_WRITE | NOTE_EXTEND | NOTE_ATTRIB | - NOTE_LINK | NOTE_RENAME | NOTE_REVOKE, 0, (void*) TASK_FD(t)); + NOTE_LINK | NOTE_RENAME | NOTE_REVOKE | addflags, 0, (void*) TASK_FD(t)); #endif if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { if (TASK_ROOT(t)->root_hooks.hook_exec.exception) @@ -866,7 +838,7 @@ static inline void fetch_hook_kevent_proceed(int en, struct kevent *res, sched_root_task_t *r) { struct kevent evt[1]; - register int i; + register int i, flg; sched_task_t *task, *tmp; struct timespec now = { 0, 0 }; #ifdef AIO_SUPPORT @@ -884,102 +856,115 @@ fetch_hook_kevent_proceed(int en, struct kevent *res, memcpy(evt, &res[i], sizeof evt); evt->flags = EV_DELETE; /* Put read/write task to ready queue */ + flg = 0; switch (res[i].filter) { case EVFILT_READ: TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { if (TASK_FD(task) == ((intptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove read handle */ - remove_task_from(task, &r->root_read); + /* remove read handle */ + remove_task_from(task, &r->root_read); - if (r->root_hooks.hook_exec.exception && res[i].flags & EV_EOF) { - if (r->root_hooks.hook_exec.exception(r, (void*) EV_EOF)) { - task->task_type = taskUNUSE; - insert_task_to(task, &r->root_unuse); + if (r->root_hooks.hook_exec.exception && res[i].flags & EV_EOF) { + if (r->root_hooks.hook_exec.exception(r, (void*) EV_EOF)) { + task->task_type = taskUNUSE; + insert_task_to(task, &r->root_unuse); + } else { + task->task_type = taskREADY; + insert_task_to(task, &r->root_ready); + } } else { task->task_type = taskREADY; insert_task_to(task, &r->root_ready); } - } else { - task->task_type = taskREADY; - insert_task_to(task, &r->root_ready); } - break; + flg++; } } break; case EVFILT_WRITE: TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { if (TASK_FD(task) == ((intptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove write handle */ - remove_task_from(task, &r->root_write); + /* remove write handle */ + remove_task_from(task, &r->root_write); - if (r->root_hooks.hook_exec.exception && res[i].flags & EV_EOF) { - if (r->root_hooks.hook_exec.exception(r, (void*) EV_EOF)) { - task->task_type = taskUNUSE; - insert_task_to(task, &r->root_unuse); + if (r->root_hooks.hook_exec.exception && res[i].flags & EV_EOF) { + if (r->root_hooks.hook_exec.exception(r, (void*) EV_EOF)) { + task->task_type = taskUNUSE; + insert_task_to(task, &r->root_unuse); + } else { + task->task_type = taskREADY; + insert_task_to(task, &r->root_ready); + } } else { task->task_type = taskREADY; insert_task_to(task, &r->root_ready); } - } else { - task->task_type = taskREADY; - insert_task_to(task, &r->root_ready); } - break; + flg++; } } break; case EVFILT_TIMER: TAILQ_FOREACH_SAFE(task, &r->root_alarm, task_node, tmp) { if ((uintptr_t) TASK_DATA(task) == ((uintptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove alarm handle */ - transit_task2ready(task, &r->root_alarm); - break; + /* remove alarm handle */ + transit_task2ready(task, &r->root_alarm); + } + flg++; } } break; case EVFILT_VNODE: TAILQ_FOREACH_SAFE(task, &r->root_node, task_node, tmp) { if (TASK_FD(task) == ((intptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove node handle */ - transit_task2ready(task, &r->root_node); - break; + /* remove node handle */ + transit_task2ready(task, &r->root_node); + } + flg++; } } break; case EVFILT_PROC: TAILQ_FOREACH_SAFE(task, &r->root_proc, task_node, tmp) { if (TASK_VAL(task) == ((uintptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove proc handle */ - transit_task2ready(task, &r->root_proc); - break; + /* remove proc handle */ + transit_task2ready(task, &r->root_proc); + } + flg++; } } break; case EVFILT_SIGNAL: TAILQ_FOREACH_SAFE(task, &r->root_signal, task_node, tmp) { if (TASK_VAL(task) == ((uintptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove signal handle */ - transit_task2ready(task, &r->root_signal); - break; + /* remove signal handle */ + transit_task2ready(task, &r->root_signal); + } + flg++; } } break; @@ -988,22 +973,24 @@ fetch_hook_kevent_proceed(int en, struct kevent *res, TAILQ_FOREACH_SAFE(task, &r->root_aio, task_node, tmp) { acb = (struct aiocb*) TASK_VAL(task); if (acb == ((struct aiocb*) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; - - /* remove user handle */ - transit_task2ready(task, &r->root_aio); + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - fd = acb->aio_fildes; - if ((len = aio_return(acb)) != -1) { - if (lseek(fd, acb->aio_offset + len, SEEK_CUR) == -1) + /* remove user handle */ + transit_task2ready(task, &r->root_aio); + + fd = acb->aio_fildes; + if ((len = aio_return(acb)) != -1) { + if (lseek(fd, acb->aio_offset + len, SEEK_CUR) == -1) + LOGERR; + } else LOGERR; - } else - LOGERR; - free(acb); - TASK_DATLEN(task) = (u_long) len; - TASK_FD(task) = fd; - break; + e_free(acb); + TASK_DATLEN(task) = (u_long) len; + TASK_FD(task) = fd; + } + flg++; } } break; @@ -1012,29 +999,31 @@ fetch_hook_kevent_proceed(int en, struct kevent *res, TAILQ_FOREACH_SAFE(task, &r->root_lio, task_node, tmp) { acbs = (struct aiocb**) TASK_VAL(task); if (acbs == ((struct aiocb**) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove user handle */ - transit_task2ready(task, &r->root_lio); + /* remove user handle */ + transit_task2ready(task, &r->root_lio); - iv = (struct iovec*) TASK_DATA(task); - fd = acbs[0]->aio_fildes; - off = acbs[0]->aio_offset; - for (len = 0; i < TASK_DATLEN(task); len += l, i++) { - if ((iv[i].iov_len = aio_return(acbs[i])) == -1) - l = 0; - else - l = iv[i].iov_len; - free(acbs[i]); - } - free(acbs); - TASK_DATLEN(task) = (u_long) len; - TASK_FD(task) = fd; + iv = (struct iovec*) TASK_DATA(task); + fd = acbs[0]->aio_fildes; + off = acbs[0]->aio_offset; + for (len = 0; i < TASK_DATLEN(task); len += l, i++) { + if ((iv[i].iov_len = aio_return(acbs[i])) == -1) + l = 0; + else + l = iv[i].iov_len; + e_free(acbs[i]); + } + e_free(acbs); + TASK_DATLEN(task) = (u_long) len; + TASK_FD(task) = fd; - if (lseek(fd, off + len, SEEK_CUR) == -1) - LOGERR; - break; + if (lseek(fd, off + len, SEEK_CUR) == -1) + LOGERR; + } + flg++; } } break; @@ -1044,18 +1033,23 @@ fetch_hook_kevent_proceed(int en, struct kevent *res, case EVFILT_USER: TAILQ_FOREACH_SAFE(task, &r->root_user, task_node, tmp) { if (TASK_VAL(task) == ((uintptr_t) res[i].udata)) { - TASK_RET(task) = res[i].data; - TASK_FLAG(task) = (u_long) res[i].fflags; + if (!flg) { + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; - /* remove user handle */ - transit_task2ready(task, &r->root_user); - break; + /* remove user handle */ + transit_task2ready(task, &r->root_user); + } + flg++; } } break; #endif /* EVFILT_USER */ } + if (flg > 1) + evt->flags &= ~EV_DELETE; + if (kevent(r->root_kq, evt, 1, NULL, 0, &now) == -1) { if (r->root_hooks.hook_exec.exception) r->root_hooks.hook_exec.exception(r, NULL); @@ -1070,26 +1064,27 @@ fetch_hook_kevent_proceed(int en, struct kevent *res, static inline void fetch_hook_epoll_proceed(int en, struct epoll_event *res, sched_root_task_t *r) { - register int i, flg; + register int i, rflg, wflg; int ops = EPOLL_CTL_DEL; sched_task_t *t, *tmp, *task; struct epoll_event evt[1]; for (i = 0; i < en; i++) { memcpy(evt, &res[i], sizeof evt); + evt->events ^= evt->events; + rflg = wflg = 0; - if (evt->events & (EPOLLIN | EPOLLPRI)) { - flg = 0; + if (res[i].events & (EPOLLIN | EPOLLPRI)) { task = NULL; TAILQ_FOREACH_SAFE(t, &r->root_read, task_node, tmp) { if (TASK_FD(t) == evt->data.fd) { - if (!flg) + if (!task) task = t; - flg++; + rflg++; } } - if (flg && task) { + if (task) { TASK_FLAG(task) = ioctl(TASK_FD(task), FIONREAD, &TASK_RET(task)); /* remove read handle */ remove_task_from(task, &r->root_read); @@ -1108,29 +1103,27 @@ fetch_hook_epoll_proceed(int en, struct epoll_event *r insert_task_to(task, &r->root_ready); } - evt->events ^= evt->events; - if (FD_ISSET(evt->data.fd, &r->root_fds[1])) { - ops = EPOLL_CTL_MOD; + if (!(res[i].events & EPOLLOUT) && FD_ISSET(evt->data.fd, &r->root_fds[1])) { evt->events |= EPOLLOUT; + wflg = 42; } - if (flg > 1) { - ops = EPOLL_CTL_MOD; + if (rflg > 1) evt->events |= EPOLLIN | EPOLLPRI; - } else + else FD_CLR(evt->data.fd, &r->root_fds[0]); } - } else if (evt->events & EPOLLOUT) { - flg = 0; + } + if (res[i].events & EPOLLOUT) { task = NULL; TAILQ_FOREACH_SAFE(t, &r->root_write, task_node, tmp) { if (TASK_FD(t) == evt->data.fd) { - if (!flg) + if (!task) task = t; - flg++; + wflg++; } } - if (flg && task) { + if (task) { TASK_FLAG(task) = ioctl(TASK_FD(task), FIONWRITE, &TASK_RET(task)); /* remove write handle */ remove_task_from(task, &r->root_write); @@ -1149,19 +1142,20 @@ fetch_hook_epoll_proceed(int en, struct epoll_event *r insert_task_to(task, &r->root_ready); } - evt->events ^= evt->events; - if (FD_ISSET(evt->data.fd, &r->root_fds[0])) { - ops = EPOLL_CTL_MOD; + if (!(res[i].events & (EPOLLIN | EPOLLPRI)) && FD_ISSET(evt->data.fd, &r->root_fds[0])) { evt->events |= EPOLLIN | EPOLLPRI; + rflg = 42; } - if (flg > 1) { - ops = EPOLL_CTL_MOD; + if (wflg > 1) evt->events |= EPOLLOUT; - } else + else FD_CLR(evt->data.fd, &r->root_fds[1]); } } + if (rflg > 1 || wflg > 1) + ops = EPOLL_CTL_MOD; + if (epoll_ctl(r->root_kq, ops, evt->data.fd, evt) == -1) { if (r->root_hooks.hook_exec.exception) { r->root_hooks.hook_exec.exception(r, NULL); @@ -1176,25 +1170,30 @@ fetch_hook_epoll_proceed(int en, struct epoll_event *r static inline void fetch_hook_select_proceed(int en, fd_set rfd, fd_set wfd, fd_set xfd, sched_root_task_t *r) { - register int i, flg; - sched_task_t *t, *tmp, *task = NULL; + register int i, rflg, wflg; + sched_task_t *t, *tmp, *task; /* skip select check if return value from select is zero */ if (!en) return; for (i = 0; i < r->root_kq; i++) { + if (!FD_ISSET(i, &r->root_fds[0]) && !FD_ISSET(i, &r->root_fds[1])) + continue; + + rflg = wflg = 0; + if (FD_ISSET(i, &rfd) || FD_ISSET(i, &xfd)) { - flg = 0; + task = NULL; TAILQ_FOREACH_SAFE(t, &r->root_read, task_node, tmp) { if (TASK_FD(t) == i) { - if (!flg) + if (!task) task = t; - flg++; + rflg++; } } - if (flg && task) { + if (task) { TASK_FLAG(task) = ioctl(TASK_FD(task), FIONREAD, &TASK_RET(task)); /* remove read handle */ @@ -1214,20 +1213,21 @@ fetch_hook_select_proceed(int en, fd_set rfd, fd_set w } /* remove resouce */ - if (flg == 1) + if (rflg == 1) FD_CLR(i, &r->root_fds[0]); } - } else if (FD_ISSET(i, &wfd)) { - flg = 0; + } + if (FD_ISSET(i, &wfd)) { + task = NULL; TAILQ_FOREACH_SAFE(t, &r->root_write, task_node, tmp) { if (TASK_FD(t) == i) { - if (!flg) + if (!task) task = t; - flg++; + wflg++; } } - if (flg && task) { + if (task) { TASK_FLAG(task) = ioctl(TASK_FD(task), FIONWRITE, &TASK_RET(task)); /* remove write handle */ @@ -1247,18 +1247,17 @@ fetch_hook_select_proceed(int en, fd_set rfd, fd_set w } /* remove resouce */ - if (flg == 1) + if (wflg == 1) FD_CLR(i, &r->root_fds[1]); } } } /* optimize select */ - for (i = r->root_kq - 1; i > 2; i--) + for (i = r->root_kq - 1; i >= 0; i--) if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) break; - if (i > 2) - r->root_kq = i + 1; + r->root_kq = i + 1; } #endif @@ -1300,39 +1299,41 @@ sched_hook_fetch(void *root, void *arg __unused) return task; } + /* if present member of task, set NOWAIT */ + if (!TAILQ_FIRST(&r->root_task)) { + /* timer tasks */ #ifdef TIMER_WITHOUT_SORT - clock_gettime(CLOCK_MONOTONIC, &now); + clock_gettime(CLOCK_MONOTONIC, &now); - sched_timespecclear(&r->root_wait); - TAILQ_FOREACH(task, &r->root_timer, task_node) { - if (!sched_timespecisset(&r->root_wait)) - r->root_wait = TASK_TS(task); - else if (sched_timespeccmp(&TASK_TS(task), &r->root_wait, -) < 0) - r->root_wait = TASK_TS(task); - } + sched_timespecclear(&r->root_wait); + TAILQ_FOREACH(task, &r->root_timer, task_node) { + if (!sched_timespecisset(&r->root_wait)) + r->root_wait = TASK_TS(task); + else if (sched_timespeccmp(&TASK_TS(task), &r->root_wait, -) < 0) + r->root_wait = TASK_TS(task); + } - if (TAILQ_FIRST(&r->root_timer)) { - m = r->root_wait; - sched_timespecsub(&m, &now, &mtmp); - r->root_wait = mtmp; - } else { - /* set wait INFTIM */ - sched_timespecinf(&r->root_wait); - } + if (TAILQ_FIRST(&r->root_timer)) { + m = r->root_wait; + sched_timespecsub(&m, &now, &mtmp); + r->root_wait = mtmp; + } else { + /* set wait INFTIM */ + sched_timespecinf(&r->root_wait); + } #else /* ! TIMER_WITHOUT_SORT */ - if (!TAILQ_FIRST(&r->root_task) && (task = TAILQ_FIRST(&r->root_timer))) { - clock_gettime(CLOCK_MONOTONIC, &now); + if ((task = TAILQ_FIRST(&r->root_timer))) { + clock_gettime(CLOCK_MONOTONIC, &now); - m = TASK_TS(task); - sched_timespecsub(&m, &now, &mtmp); - r->root_wait = mtmp; - } else { - /* set wait INFTIM */ - sched_timespecinf(&r->root_wait); - } + m = TASK_TS(task); + sched_timespecsub(&m, &now, &mtmp); + r->root_wait = mtmp; + } else { + /* set wait INFTIM */ + sched_timespecinf(&r->root_wait); + } #endif /* TIMER_WITHOUT_SORT */ - /* if present member of task, set NOWAIT */ - if (TAILQ_FIRST(&r->root_task)) + } else /* no waiting for event, because we have ready task */ sched_timespecclear(&r->root_wait); if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) {