--- libaitsched/src/hooks.c 2012/05/31 22:31:48 1.8 +++ libaitsched/src/hooks.c 2014/01/28 16:58:33 1.26 @@ -3,7 +3,7 @@ * by Michael Pounov * * $Author: misho $ -* $Id: hooks.c,v 1.8 2012/05/31 22:31:48 misho Exp $ +* $Id: hooks.c,v 1.26 2014/01/28 16:58:33 misho Exp $ * ************************************************************************** The ELWIX and AITNET software is distributed under the following @@ -12,7 +12,7 @@ terms: All of the documentation and software included in the ELWIX and AITNET Releases is copyrighted by ELWIX - Sofia/Bulgaria -Copyright 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012 +Copyright 2004 - 2014 by Michael Pounov . All rights reserved. Redistribution and use in source and binary forms, with or without @@ -62,11 +62,17 @@ sched_hook_init(void *root, void *arg __unused) if (!r) return (void*) -1; +#ifndef KQ_DISABLE r->root_kq = kqueue(); if (r->root_kq == -1) { LOGERR; return (void*) -1; } +#else + r->root_kq ^= r->root_kq; + FD_ZERO(&r->root_fds[0]); + FD_ZERO(&r->root_fds[1]); +#endif return NULL; } @@ -86,10 +92,16 @@ sched_hook_fini(void *root, void *arg __unused) if (!r) return (void*) -1; +#ifndef KQ_DISABLE if (r->root_kq > 2) { close(r->root_kq); r->root_kq = 0; } +#else + FD_ZERO(&r->root_fds[1]); + FD_ZERO(&r->root_fds[0]); + r->root_kq ^= r->root_kq; +#endif return NULL; } @@ -105,28 +117,66 @@ void * sched_hook_cancel(void *task, void *arg __unused) { sched_task_t *t = task; +#ifndef KQ_DISABLE struct kevent chg[1]; struct timespec timeout = { 0, 0 }; +#else + sched_root_task_t *r = NULL; + register int i; +#endif +#ifdef AIO_SUPPORT + struct aiocb *acb; +#ifdef EVFILT_LIO + register int i = 0; + struct aiocb **acbs; +#endif /* EVFILT_LIO */ +#endif /* AIO_SUPPORT */ if (!t || !TASK_ROOT(t)) return (void*) -1; +#ifdef KQ_DISABLE + r = TASK_ROOT(t); +#endif switch (TASK_TYPE(t)) { case taskREAD: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); #else EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); #endif +#else + FD_CLR(TASK_FD(t), &r->root_fds[0]); + + /* optimize select */ + for (i = r->root_kq - 1; i > 2; i--) + if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) + break; + if (i > 2) + r->root_kq = i + 1; +#endif break; case taskWRITE: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); #else EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); #endif +#else + FD_CLR(TASK_FD(t), &r->root_fds[1]); + + /* optimize select */ + for (i = r->root_kq - 1; i > 2; i--) + if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) + break; + if (i > 2) + r->root_kq = i + 1; +#endif break; case taskALARM: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, 0, 0, (intptr_t) TASK_DATA(t)); @@ -134,30 +184,79 @@ sched_hook_cancel(void *task, void *arg __unused) EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_DELETE, 0, 0, (void*) TASK_DATA(t)); #endif +#endif break; case taskNODE: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); #else EV_SET(&chg[0], TASK_FD(t), EVFILT_VNODE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); #endif +#endif break; case taskPROC: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); #else EV_SET(&chg[0], TASK_VAL(t), EVFILT_PROC, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); #endif +#endif break; case taskSIGNAL: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); #else EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); #endif + /* restore signal */ + signal(TASK_VAL(t), SIG_DFL); +#endif break; +#ifdef AIO_SUPPORT + case taskAIO: +#ifndef KQ_DISABLE +#ifdef __NetBSD__ + EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); +#else + EV_SET(&chg[0], TASK_VAL(t), EVFILT_AIO, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); +#endif + acb = (struct aiocb*) TASK_VAL(t); + if (acb) { + if (aio_cancel(acb->aio_fildes, acb) == AIO_CANCELED) + aio_return(acb); + free(acb); + TASK_VAL(t) = 0; + } +#endif + break; +#ifdef EVFILT_LIO + case taskLIO: +#ifndef KQ_DISABLE +#ifdef __NetBSD__ + EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); +#else + EV_SET(&chg[0], TASK_VAL(t), EVFILT_LIO, EV_DELETE, 0, 0, (void*) TASK_VAL(t)); +#endif + acbs = (struct aiocb**) TASK_VAL(t); + if (acbs) { + for (i = 0; i < TASK_DATLEN(t); i++) { + if (aio_cancel(acbs[i]->aio_fildes, acbs[i]) == AIO_CANCELED) + aio_return(acbs[i]); + free(acbs[i]); + } + free(acbs); + TASK_VAL(t) = 0; + } +#endif + break; +#endif /* EVFILT_LIO */ +#endif /* AIO_SUPPORT */ #ifdef EVFILT_USER case taskUSER: +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_USER, EV_DELETE, 0, 0, (intptr_t) TASK_VAL(t)); #else @@ -165,15 +264,65 @@ sched_hook_cancel(void *task, void *arg __unused) #endif #endif break; +#endif /* EVFILT_USER */ + case taskTHREAD: +#ifdef HAVE_LIBPTHREAD + pthread_cancel((pthread_t) TASK_VAL(t)); +#endif + return NULL; +#if defined(HAVE_TIMER_CREATE) && defined(HAVE_TIMER_SETTIME) + case taskRTC: + timer_delete((timer_t) TASK_FLAG(t)); + schedCancel((sched_task_t*) TASK_RET(t)); + return NULL; +#endif /* HAVE_TIMER_CREATE */ default: return NULL; } +#ifndef KQ_DISABLE kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); +#endif return NULL; } +#ifdef HAVE_LIBPTHREAD /* + * sched_hook_thread() - Default THREAD hook + * + * @task = current task + * @arg = pthread attributes + * return: <0 errors and 0 ok + */ +void * +sched_hook_thread(void *task, void *arg) +{ + sched_task_t *t = task; + pthread_t tid; + sigset_t s, o; + + if (!t || !TASK_ROOT(t)) + return (void*) -1; + + sigfillset(&s); + pthread_sigmask(SIG_BLOCK, &s, &o); + if ((errno = pthread_create(&tid, (pthread_attr_t*) arg, + (void *(*)(void*)) _sched_threadWrapper, t))) { + LOGERR; + pthread_sigmask(SIG_SETMASK, &o, NULL); + return (void*) -1; + } else + TASK_VAL(t) = (u_long) tid; + + if (!TASK_ISLOCKED(t)) + TASK_LOCK(t); + + pthread_sigmask(SIG_SETMASK, &o, NULL); + return NULL; +} +#endif + +/* * sched_hook_read() - Default READ hook * * @task = current task @@ -184,12 +333,20 @@ void * sched_hook_read(void *task, void *arg __unused) { sched_task_t *t = task; +#ifndef KQ_DISABLE struct kevent chg[1]; struct timespec timeout = { 0, 0 }; +#else + sched_root_task_t *r = NULL; +#endif if (!t || !TASK_ROOT(t)) return (void*) -1; +#ifdef KQ_DISABLE + r = TASK_ROOT(t); +#endif +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); #else @@ -202,6 +359,11 @@ sched_hook_read(void *task, void *arg __unused) LOGERR; return (void*) -1; } +#else + FD_SET(TASK_FD(t), &r->root_fds[0]); + if (TASK_FD(t) >= r->root_kq) + r->root_kq = TASK_FD(t) + 1; +#endif return NULL; } @@ -217,12 +379,20 @@ void * sched_hook_write(void *task, void *arg __unused) { sched_task_t *t = task; +#ifndef KQ_DISABLE struct kevent chg[1]; struct timespec timeout = { 0, 0 }; +#else + sched_root_task_t *r = NULL; +#endif if (!t || !TASK_ROOT(t)) return (void*) -1; +#ifdef KQ_DISABLE + r = TASK_ROOT(t); +#endif +#ifndef KQ_DISABLE #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); #else @@ -235,6 +405,11 @@ sched_hook_write(void *task, void *arg __unused) LOGERR; return (void*) -1; } +#else + FD_SET(TASK_FD(t), &r->root_fds[1]); + if (TASK_FD(t) >= r->root_kq) + r->root_kq = TASK_FD(t) + 1; +#endif return NULL; } @@ -249,6 +424,7 @@ sched_hook_write(void *task, void *arg __unused) void * sched_hook_alarm(void *task, void *arg __unused) { +#ifndef KQ_DISABLE sched_task_t *t = task; struct kevent chg[1]; struct timespec timeout = { 0, 0 }; @@ -257,11 +433,11 @@ sched_hook_alarm(void *task, void *arg __unused) return (void*) -1; #ifdef __NetBSD__ - EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, + EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_CLEAR, 0, t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, (intptr_t) TASK_DATA(t)); #else - EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_ONESHOT, 0, + EV_SET(&chg[0], (uintptr_t) TASK_DATA(t), EVFILT_TIMER, EV_ADD | EV_CLEAR, 0, t->task_val.ts.tv_sec * 1000 + t->task_val.ts.tv_nsec / 1000000, (void*) TASK_DATA(t)); #endif @@ -273,6 +449,7 @@ sched_hook_alarm(void *task, void *arg __unused) return (void*) -1; } +#endif return NULL; } @@ -286,6 +463,7 @@ sched_hook_alarm(void *task, void *arg __unused) void * sched_hook_node(void *task, void *arg __unused) { +#ifndef KQ_DISABLE sched_task_t *t = task; struct kevent chg[1]; struct timespec timeout = { 0, 0 }; @@ -310,6 +488,7 @@ sched_hook_node(void *task, void *arg __unused) return (void*) -1; } +#endif return NULL; } @@ -323,6 +502,7 @@ sched_hook_node(void *task, void *arg __unused) void * sched_hook_proc(void *task, void *arg __unused) { +#ifndef KQ_DISABLE sched_task_t *t = task; struct kevent chg[1]; struct timespec timeout = { 0, 0 }; @@ -345,6 +525,7 @@ sched_hook_proc(void *task, void *arg __unused) return (void*) -1; } +#endif return NULL; } @@ -358,6 +539,7 @@ sched_hook_proc(void *task, void *arg __unused) void * sched_hook_signal(void *task, void *arg __unused) { +#ifndef KQ_DISABLE sched_task_t *t = task; struct kevent chg[1]; struct timespec timeout = { 0, 0 }; @@ -365,10 +547,13 @@ sched_hook_signal(void *task, void *arg __unused) if (!t || !TASK_ROOT(t)) return (void*) -1; + /* ignore signal */ + signal(TASK_VAL(t), SIG_IGN); + #ifdef __NetBSD__ - EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (intptr_t) TASK_VAL(t)); + EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_VAL(t)); #else - EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD, 0, 0, (void*) TASK_VAL(t)); + EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_VAL(t)); #endif if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { if (TASK_ROOT(t)->root_hooks.hook_exec.exception) @@ -377,7 +562,25 @@ sched_hook_signal(void *task, void *arg __unused) LOGERR; return (void*) -1; } +#else +#if 0 + sched_task_t *t = task; + struct sigaction sa; + memset(&sa, 0, sizeof sa); + sigemptyset(&sa.sa_mask); + sa.sa_handler = _sched_sigHandler; + sa.sa_flags = SA_RESETHAND | SA_RESTART; + + if (sigaction(TASK_VAL(t), &sa, NULL) == -1) { + if (TASK_ROOT(t)->root_hooks.hook_exec.exception) + TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); + else + LOGERR; + return (void*) -1; + } +#endif /* 0 */ +#endif return NULL; } @@ -392,6 +595,7 @@ sched_hook_signal(void *task, void *arg __unused) void * sched_hook_user(void *task, void *arg __unused) { +#ifndef KQ_DISABLE sched_task_t *t = task; struct kevent chg[1]; struct timespec timeout = { 0, 0 }; @@ -414,6 +618,7 @@ sched_hook_user(void *task, void *arg __unused) return (void*) -1; } +#endif return NULL; } #endif @@ -431,10 +636,26 @@ sched_hook_fetch(void *root, void *arg __unused) sched_root_task_t *r = root; sched_task_t *task, *tmp; struct timespec now, m, mtmp; - struct timespec *timeout; +#ifndef KQ_DISABLE struct kevent evt[1], res[KQ_EVENTS]; - register int i; + struct timespec *timeout; +#else + struct timeval *timeout, tv; + fd_set rfd, wfd, xfd; +#endif + register int i, flg; int en; +#ifdef AIO_SUPPORT + int len, fd; + struct aiocb *acb; +#ifdef EVFILT_LIO + int l; + register int j; + off_t off; + struct aiocb **acbs; + struct iovec *iv; +#endif /* EVFILT_LIO */ +#endif /* AIO_SUPPORT */ if (!r) return NULL; @@ -496,8 +717,8 @@ sched_hook_fetch(void *root, void *arg __unused) /* set wait INFTIM */ sched_timespecinf(&r->root_wait); } -#else - if (!TAILQ_FIRST(&r->root_eventlo) && (task = TAILQ_FIRST(&r->root_timer))) { +#else /* ! TIMER_WITHOUT_SORT */ + if (!TAILQ_FIRST(&r->root_task) && (task = TAILQ_FIRST(&r->root_timer))) { clock_gettime(CLOCK_MONOTONIC, &now); m = TASK_TS(task); @@ -507,37 +728,63 @@ sched_hook_fetch(void *root, void *arg __unused) /* set wait INFTIM */ sched_timespecinf(&r->root_wait); } -#endif - /* if present member of eventLo, set NOWAIT */ - if (TAILQ_FIRST(&r->root_eventlo)) +#endif /* TIMER_WITHOUT_SORT */ + /* if present member of task, set NOWAIT */ + if (TAILQ_FIRST(&r->root_task)) sched_timespecclear(&r->root_wait); - if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) + if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) { +#ifndef KQ_DISABLE timeout = &r->root_wait; - else if (sched_timespecisinf(&r->root_poll)) +#else + sched_timespec2val(&r->root_wait, &tv); + timeout = &tv; +#endif /* KQ_DISABLE */ + } else if (sched_timespecisinf(&r->root_poll)) timeout = NULL; - else + else { +#ifndef KQ_DISABLE timeout = &r->root_poll; +#else + sched_timespec2val(&r->root_poll, &tv); + timeout = &tv; +#endif /* KQ_DISABLE */ + } + +#ifndef KQ_DISABLE if ((en = kevent(r->root_kq, NULL, 0, res, KQ_EVENTS, timeout)) == -1) { +#else + rfd = xfd = r->root_fds[0]; + wfd = r->root_fds[1]; + if ((en = select(r->root_kq, &rfd, &wfd, &xfd, timeout)) == -1) { +#endif /* KQ_DISABLE */ if (r->root_hooks.hook_exec.exception) { if (r->root_hooks.hook_exec.exception(r, NULL)) return NULL; } else if (errno != EINTR) LOGERR; - return NULL; + goto skip_event; } + /* kevent dispatcher */ now.tv_sec = now.tv_nsec = 0; /* Go and catch the cat into pipes ... */ +#ifndef KQ_DISABLE for (i = 0; i < en; i++) { memcpy(evt, &res[i], sizeof evt); evt->flags = EV_DELETE; /* Put read/write task to ready queue */ switch (res[i].filter) { case EVFILT_READ: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { if (TASK_FD(task) != ((intptr_t) res[i].udata)) continue; + else { + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; + } /* remove read handle */ #ifdef HAVE_LIBPTHREAD pthread_mutex_lock(&r->root_mtx[taskREAD]); @@ -576,13 +823,21 @@ sched_hook_fetch(void *root, void *arg __unused) pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif } - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; case EVFILT_WRITE: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { if (TASK_FD(task) != ((intptr_t) res[i].udata)) continue; + else { + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; + } /* remove write handle */ #ifdef HAVE_LIBPTHREAD pthread_mutex_lock(&r->root_mtx[taskWRITE]); @@ -621,13 +876,21 @@ sched_hook_fetch(void *root, void *arg __unused) pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif } - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; case EVFILT_TIMER: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_alarm, task_node, tmp) { if ((uintptr_t) TASK_DATA(task) != ((uintptr_t) res[i].udata)) continue; + else { + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; + } /* remove alarm handle */ #ifdef HAVE_LIBPTHREAD pthread_mutex_lock(&r->root_mtx[taskALARM]); @@ -644,16 +907,20 @@ sched_hook_fetch(void *root, void *arg __unused) #ifdef HAVE_LIBPTHREAD pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; case EVFILT_VNODE: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_node, task_node, tmp) { if (TASK_FD(task) != ((intptr_t) res[i].udata)) continue; else { - TASK_DATA(task) = (void*) (uintptr_t) res[i].data; - TASK_DATLEN(task) = res[i].fflags; + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; } /* remove node handle */ #ifdef HAVE_LIBPTHREAD @@ -671,16 +938,20 @@ sched_hook_fetch(void *root, void *arg __unused) #ifdef HAVE_LIBPTHREAD pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; case EVFILT_PROC: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_proc, task_node, tmp) { if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) continue; else { - TASK_DATA(task) = (void*) (uintptr_t) res[i].data; - TASK_DATLEN(task) = res[i].fflags; + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; } /* remove proc handle */ #ifdef HAVE_LIBPTHREAD @@ -698,13 +969,21 @@ sched_hook_fetch(void *root, void *arg __unused) #ifdef HAVE_LIBPTHREAD pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; case EVFILT_SIGNAL: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_signal, task_node, tmp) { if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) continue; + else { + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; + } /* remove signal handle */ #ifdef HAVE_LIBPTHREAD pthread_mutex_lock(&r->root_mtx[taskSIGNAL]); @@ -721,17 +1000,114 @@ sched_hook_fetch(void *root, void *arg __unused) #ifdef HAVE_LIBPTHREAD pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; +#ifdef AIO_SUPPORT + case EVFILT_AIO: + flg = 0; + TAILQ_FOREACH_SAFE(task, &r->root_aio, task_node, tmp) { + acb = (struct aiocb*) TASK_VAL(task); + if (acb != ((struct aiocb*) res[i].udata)) + continue; + else { + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; + } + /* remove user handle */ +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskAIO]); +#endif + TAILQ_REMOVE(&r->root_aio, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskAIO]); +#endif + task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif + TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif + fd = acb->aio_fildes; + if ((len = aio_return(acb)) != -1) { + if (lseek(fd, acb->aio_offset + len, SEEK_CUR) == -1) + LOGERR; + } else + LOGERR; + free(acb); + TASK_DATLEN(task) = (u_long) len; + TASK_FD(task) = fd; + } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; + break; +#ifdef EVFILT_LIO + case EVFILT_LIO: + flg = 0; + TAILQ_FOREACH_SAFE(task, &r->root_lio, task_node, tmp) { + acbs = (struct aiocb**) TASK_VAL(task); + if (acbs != ((struct aiocb**) res[i].udata)) + continue; + else { + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; + } + /* remove user handle */ +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskLIO]); +#endif + TAILQ_REMOVE(&r->root_lio, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskLIO]); +#endif + task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif + TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif + iv = (struct iovec*) TASK_DATA(task); + fd = acbs[0]->aio_fildes; + off = acbs[0]->aio_offset; + for (j = len = 0; i < TASK_DATLEN(task); len += l, i++) { + if ((iv[i].iov_len = aio_return(acbs[i])) == -1) + l = 0; + else + l = iv[i].iov_len; + free(acbs[i]); + } + free(acbs); + TASK_DATLEN(task) = (u_long) len; + TASK_FD(task) = fd; + + if (lseek(fd, off + len, SEEK_CUR) == -1) + LOGERR; + } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; + break; +#endif /* EVFILT_LIO */ +#endif /* AIO_SUPPORT */ #ifdef EVFILT_USER case EVFILT_USER: + flg = 0; TAILQ_FOREACH_SAFE(task, &r->root_user, task_node, tmp) { if (TASK_VAL(task) != ((uintptr_t) res[i].udata)) continue; else { - TASK_DATA(task) = (void*) res[i].data; - TASK_DATLEN(task) = res[i].fflags; + flg++; + TASK_RET(task) = res[i].data; + TASK_FLAG(task) = (u_long) res[i].fflags; } /* remove user handle */ #ifdef HAVE_LIBPTHREAD @@ -749,10 +1125,12 @@ sched_hook_fetch(void *root, void *arg __unused) #ifdef HAVE_LIBPTHREAD pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif - break; } + /* if match at least 2, don't remove resouce of event */ + if (flg > 1) + evt->flags ^= evt->flags; break; -#endif +#endif /* EVFILT_USER */ } if (kevent(r->root_kq, evt, 1, NULL, 0, &now) == -1) { if (r->root_hooks.hook_exec.exception) { @@ -762,7 +1140,126 @@ sched_hook_fetch(void *root, void *arg __unused) LOGERR; } } +#else /* end of kevent dispatcher */ + for (i = 0; i < r->root_kq; i++) { + if (FD_ISSET(i, &rfd) || FD_ISSET(i, &xfd)) { + flg = 0; + TAILQ_FOREACH_SAFE(task, &r->root_read, task_node, tmp) { + if (TASK_FD(task) != i) + continue; + else { + flg++; + TASK_FLAG(task) = ioctl(TASK_FD(task), + FIONREAD, &TASK_RET(task)); + } + /* remove read handle */ +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREAD]); +#endif + TAILQ_REMOVE(&r->root_read, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREAD]); +#endif + if (r->root_hooks.hook_exec.exception) { + if (r->root_hooks.hook_exec.exception(r, NULL)) { + task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif + TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif + } else { + task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif + TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif + } + } else { + task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif + TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif + } + } + /* if match equal to 1, remove resouce */ + if (flg == 1) + FD_CLR(i, &r->root_fds[0]); + } + if (FD_ISSET(i, &wfd)) { + flg = 0; + TAILQ_FOREACH_SAFE(task, &r->root_write, task_node, tmp) { + if (TASK_FD(task) != i) + continue; + else { + flg++; + TASK_FLAG(task) = ioctl(TASK_FD(task), + FIONWRITE, &TASK_RET(task)); + } + /* remove write handle */ +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskWRITE]); +#endif + TAILQ_REMOVE(&r->root_write, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskWRITE]); +#endif + if (r->root_hooks.hook_exec.exception) { + if (r->root_hooks.hook_exec.exception(r, NULL)) { + task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif + TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif + } else { + task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif + TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif + } + } else { + task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif + TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif + } + } + /* if match equal to 1, remove resouce */ + if (flg == 1) + FD_CLR(i, &r->root_fds[1]); + } + } + + /* optimize select */ + for (i = r->root_kq - 1; i > 2; i--) + if (FD_ISSET(i, &r->root_fds[0]) || FD_ISSET(i, &r->root_fds[1])) + break; + if (i > 2) + r->root_kq = i + 1; +#endif /* KQ_DISABLE */ + +skip_event: /* timer update & put in ready queue */ clock_gettime(CLOCK_MONOTONIC, &now); @@ -785,18 +1282,18 @@ sched_hook_fetch(void *root, void *arg __unused) #endif } - /* put eventlo priority task to ready queue, if there is no ready task or - reach max missed fetch-rotate */ - if ((task = TAILQ_FIRST(&r->root_eventlo))) { - if (!TAILQ_FIRST(&r->root_ready) || r->root_eventlo_miss > MAX_EVENTLO_MISS) { - r->root_eventlo_miss = 0; + /* put regular task priority task to ready queue, + if there is no ready task or reach max missing hit for regular task */ + if ((task = TAILQ_FIRST(&r->root_task))) { + if (!TAILQ_FIRST(&r->root_ready) || r->root_miss >= TASK_VAL(task)) { + r->root_miss ^= r->root_miss; #ifdef HAVE_LIBPTHREAD - pthread_mutex_lock(&r->root_mtx[taskEVENTLO]); + pthread_mutex_lock(&r->root_mtx[taskTASK]); #endif - TAILQ_REMOVE(&r->root_eventlo, task, task_node); + TAILQ_REMOVE(&r->root_task, task, task_node); #ifdef HAVE_LIBPTHREAD - pthread_mutex_unlock(&r->root_mtx[taskEVENTLO]); + pthread_mutex_unlock(&r->root_mtx[taskTASK]); #endif task->task_type = taskREADY; #ifdef HAVE_LIBPTHREAD @@ -807,9 +1304,9 @@ sched_hook_fetch(void *root, void *arg __unused) pthread_mutex_unlock(&r->root_mtx[taskREADY]); #endif } else - r->root_eventlo_miss++; + r->root_miss++; } else - r->root_eventlo_miss = 0; + r->root_miss ^= r->root_miss; /* OK, lets get ready task !!! */ task = TAILQ_FIRST(&r->root_ready); @@ -882,3 +1379,65 @@ sched_hook_condition(void *root, void *arg) return (void*) (r->root_cond - *(intptr_t*) arg); } + +/* + * sched_hook_rtc() - Default RTC hook + * + * @task = current task + * @arg = unused + * return: <0 errors and 0 ok + */ +#if defined(HAVE_TIMER_CREATE) && defined(HAVE_TIMER_SETTIME) +void * +sched_hook_rtc(void *task, void *arg __unused) +{ + sched_task_t *sigt = NULL, *t = task; + struct itimerspec its; + struct sigevent evt; + timer_t tmr; + + if (!t || !TASK_ROOT(t)) + return (void*) -1; + + memset(&evt, 0, sizeof evt); + evt.sigev_notify = SIGEV_SIGNAL; + evt.sigev_signo = (intptr_t) TASK_DATA(t) + SIGRTMIN; + evt.sigev_value.sival_ptr = TASK_DATA(t); + + if (timer_create(CLOCK_MONOTONIC, &evt, &tmr) == -1) { + if (TASK_ROOT(t)->root_hooks.hook_exec.exception) + TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); + else + LOGERR; + return (void*) -1; + } else + TASK_FLAG(t) = (u_long) tmr; + + if (!(sigt = schedSignal(TASK_ROOT(t), _sched_rtcWrapper, TASK_ARG(t), evt.sigev_signo, + t, (size_t) tmr))) { + if (TASK_ROOT(t)->root_hooks.hook_exec.exception) + TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); + else + LOGERR; + timer_delete(tmr); + return (void*) -1; + } else + TASK_RET(t) = (uintptr_t) sigt; + + memset(&its, 0, sizeof its); + its.it_value.tv_sec = t->task_val.ts.tv_sec; + its.it_value.tv_nsec = t->task_val.ts.tv_nsec; + + if (timer_settime(tmr, TIMER_RELTIME, &its, NULL) == -1) { + if (TASK_ROOT(t)->root_hooks.hook_exec.exception) + TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); + else + LOGERR; + schedCancel(sigt); + timer_delete(tmr); + return (void*) -1; + } + + return NULL; +} +#endif /* HAVE_TIMER_CREATE */