--- libaitsched/src/hooks.c 2011/10/04 13:50:05 1.2.2.2 +++ libaitsched/src/hooks.c 2012/01/08 03:28:26 1.3.4.1 @@ -3,7 +3,7 @@ * by Michael Pounov * * $Author: misho $ -* $Id: hooks.c,v 1.2.2.2 2011/10/04 13:50:05 misho Exp $ +* $Id: hooks.c,v 1.3.4.1 2012/01/08 03:28:26 misho Exp $ * ************************************************************************** The ELWIX and AITNET software is distributed under the following @@ -121,19 +121,19 @@ sched_hook_cancel(void *task, void *arg __unused) struct kevent chg[1]; struct timespec timeout = { 0, 0 }; - if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) + if (!t || !TASK_ROOT(t) || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) return (void*) -1; else io = ROOT_DATA(t->task_root); - switch (t->task_type) { + switch (TASK_TYPE(t)) { case taskREAD: #ifdef __NetBSD__ EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); #else EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); #endif - kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout); + kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); FD_CLR(TASK_FD(t), &io->rfd); break; case taskWRITE: @@ -142,7 +142,7 @@ sched_hook_cancel(void *task, void *arg __unused) #else EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); #endif - kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout); + kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout); FD_CLR(TASK_FD(t), &io->wfd); break; default: @@ -166,7 +166,7 @@ sched_hook_read(void *task, void *arg __unused) struct kevent chg[1]; struct timespec timeout = { 0, 0 }; - if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) + if (!t || !TASK_ROOT(t) || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) return (void*) -1; else io = ROOT_DATA(t->task_root); @@ -181,9 +181,9 @@ sched_hook_read(void *task, void *arg __unused) #else EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (void*) TASK_FD(t)); #endif - if (kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout) == -1) { - if (t->task_root->root_hooks.hook_exec.exception) - t->task_root->root_hooks.hook_exec.exception(t->task_root, NULL); + if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { + if (TASK_ROOT(t)->root_hooks.hook_exec.exception) + TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); else LOGERR; return (void*) -1; @@ -206,7 +206,7 @@ sched_hook_write(void *task, void *arg __unused) struct kevent chg[1]; struct timespec timeout = { 0, 0 }; - if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) + if (!t || !TASK_ROOT(t) || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) return (void*) -1; else io = ROOT_DATA(t->task_root); @@ -221,9 +221,9 @@ sched_hook_write(void *task, void *arg __unused) #else EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (void*) TASK_FD(t)); #endif - if (kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout) == -1) { - if (t->task_root->root_hooks.hook_exec.exception) - t->task_root->root_hooks.hook_exec.exception(t->task_root, NULL); + if (kevent(TASK_ROOT(t)->root_kq, chg, 1, NULL, 0, &timeout) == -1) { + if (TASK_ROOT(t)->root_hooks.hook_exec.exception) + TASK_ROOT(t)->root_hooks.hook_exec.exception(TASK_ROOT(t), NULL); else LOGERR; return (void*) -1; @@ -256,15 +256,39 @@ sched_hook_fetch(void *root, void *arg __unused) /* get new task by queue priority */ retry: while ((task = TAILQ_FIRST(&r->root_event))) { +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskEVENT]); +#endif TAILQ_REMOVE(&r->root_event, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskEVENT]); +#endif task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif return task; } while ((task = TAILQ_FIRST(&r->root_ready))) { +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_REMOVE(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif return task; } @@ -316,7 +340,7 @@ retry: if ((en = kevent(r->root_kq, NULL, 0, res, KQ_EVENTS, timeout)) == -1) { if (r->root_hooks.hook_exec.exception) { if (r->root_hooks.hook_exec.exception(r, NULL)) - return NULL; /* exit from scheduler */ + return NULL; } else LOGERR; #ifdef NDEBUG @@ -337,48 +361,96 @@ retry: switch (res[i].filter) { case EVFILT_READ: TAILQ_FOREACH(task, &r->root_read, task_node) { - if (TASK_FD(task) != ((int) res[i].udata)) + if (TASK_FD(task) != ((intptr_t) res[i].udata)) continue; /* remove read handle */ io = ROOT_DATA(task->task_root); FD_CLR(TASK_FD(task), &io->rfd); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREAD]); +#endif TAILQ_REMOVE(&r->root_read, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREAD]); +#endif if (r->root_hooks.hook_exec.exception && res[i].flags & EV_EOF) { if (r->root_hooks.hook_exec.exception(r, (void*) EV_EOF)) { task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif } else { task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif } } else { task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif } break; } break; case EVFILT_WRITE: TAILQ_FOREACH(task, &r->root_write, task_node) { - if (TASK_FD(task) != ((int) res[i].udata)) + if (TASK_FD(task) != ((intptr_t) res[i].udata)) continue; /* remove write handle */ io = ROOT_DATA(task->task_root); FD_CLR(TASK_FD(task), &io->wfd); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskWRITE]); +#endif TAILQ_REMOVE(&r->root_write, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskWRITE]); +#endif if (r->root_hooks.hook_exec.exception && res[i].flags & EV_EOF) { if (r->root_hooks.hook_exec.exception(r, (void*) EV_EOF)) { task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif } else { task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif } } else { task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif } break; } @@ -387,7 +459,7 @@ retry: if (kevent(r->root_kq, evt, 1, NULL, 0, &nw) == -1) { if (r->root_hooks.hook_exec.exception) { if (r->root_hooks.hook_exec.exception(r, NULL)) - return NULL; /* exit from scheduler */ + return NULL; } else LOGERR; } @@ -400,9 +472,21 @@ retry: TAILQ_FOREACH(task, &r->root_timer, task_node) if (timercmp(&now, &TASK_TV(task), -) >= 0) { +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskTIMER]); +#endif TAILQ_REMOVE(&r->root_timer, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskTIMER]); +#endif task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif } /* put eventlo priority task to ready queue, if there is no ready task or @@ -411,9 +495,21 @@ retry: if (!TAILQ_FIRST(&r->root_ready) || r->root_eventlo_miss > MAX_EVENTLO_MISS) { r->root_eventlo_miss = 0; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskEVENTLO]); +#endif TAILQ_REMOVE(&r->root_eventlo, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskEVENTLO]); +#endif task->task_type = taskREADY; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif } else r->root_eventlo_miss++; } else @@ -422,9 +518,21 @@ retry: /* OK, lets get ready task !!! */ if (!(task = TAILQ_FIRST(&r->root_ready))) goto retry; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskREADY]); +#endif TAILQ_REMOVE(&r->root_ready, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskREADY]); +#endif task->task_type = taskUNUSE; +#ifdef HAVE_LIBPTHREAD + pthread_mutex_lock(&r->root_mtx[taskUNUSE]); +#endif TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); +#ifdef HAVE_LIBPTHREAD + pthread_mutex_unlock(&r->root_mtx[taskUNUSE]); +#endif return task; } @@ -451,7 +559,7 @@ sched_hook_exception(void *root, void *arg) /* if error hook exists */ if (r->root_hooks.hook_root.error) - return (r->root_hooks.hook_root.error(root, (void*) errno)); + return (r->root_hooks.hook_root.error(root, (void*) ((intptr_t) errno))); /* default case! */ LOGERR;