--- libaitsched/src/hooks.c 2022/11/28 23:48:33 1.37.2.3 +++ libaitsched/src/hooks.c 2023/08/17 14:13:07 1.41.4.1 @@ -3,7 +3,7 @@ * by Michael Pounov * * $Author: misho $ -* $Id: hooks.c,v 1.37.2.3 2022/11/28 23:48:33 misho Exp $ +* $Id: hooks.c,v 1.41.4.1 2023/08/17 14:13:07 misho Exp $ * ************************************************************************** The ELWIX and AITNET software is distributed under the following @@ -12,7 +12,7 @@ terms: All of the documentation and software included in the ELWIX and AITNET Releases is copyrighted by ELWIX - Sofia/Bulgaria -Copyright 2004 - 2022 +Copyright 2004 - 2023 by Michael Pounov . All rights reserved. Redistribution and use in source and binary forms, with or without @@ -47,15 +47,6 @@ SUCH DAMAGE. #include "hooks.h" -static inline void -transit_task2ready(sched_task_t * __restrict t, sched_queue_t * __restrict q) -{ - remove_task_from(t, q); - - t->task_type = taskREADY; - insert_task_to(t, &(TASK_ROOT(t))->root_ready); -} - #ifdef HAVE_LIBPTHREAD static void * _sched_threadWrapper(sched_task_t *t) @@ -213,7 +204,7 @@ sched_hook_cancel(void *task, void *arg __unused) struct kevent chg[1]; struct timespec timeout = { 0, 0 }; #elif SUP_ENABLE == EP_SUPPORT - struct epoll_event ee = { .events = 0, .data.fd = 0 }; + struct epoll_event ee = { .events = 0, .data.u64 = 0l }; #else register int i; #endif @@ -359,11 +350,11 @@ sched_hook_cancel(void *task, void *arg __unused) #endif break; case taskSIGNAL: -#if SUP_ENABLE == KQ_SUPPORT /* check for multi subscribers */ TAILQ_FOREACH_SAFE(tt, &r->root_signal, task_node, tmp) if (TASK_VAL(tt) == TASK_VAL(t)) flg++; +#if SUP_ENABLE == KQ_SUPPORT #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, flg < 2 ? EV_DELETE : 0, 0, 0, (intptr_t) TASK_VAL(t)); @@ -371,10 +362,12 @@ sched_hook_cancel(void *task, void *arg __unused) EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, flg < 2 ? EV_DELETE : 0, 0, 0, (void*) TASK_VAL(t)); #endif - /* restore signal */ - if (flg < 2) - signal(TASK_VAL(t), SIG_DFL); #endif + if (flg < 2) { + pthread_mutex_lock(&TASK_ROOT(t)->root_sigmtx); + sigdelset(&TASK_ROOT(t)->root_sigset, TASK_VAL(t)); + pthread_mutex_unlock(&TASK_ROOT(t)->root_sigmtx); + } break; #ifdef AIO_SUPPORT case taskAIO: @@ -542,7 +535,7 @@ sched_hook_read(void *task, void *arg) struct kevent chg[1]; struct timespec timeout = { 0, 0 }; #elif SUP_ENABLE == EP_SUPPORT - struct epoll_event ee; + struct epoll_event ee = { 0 }; int flg = 0; #endif @@ -622,15 +615,16 @@ sched_hook_read(void *task, void *arg) * return: <0 errors and 0 ok */ void * -sched_hook_write(void *task, void *arg __unused) +sched_hook_write(void *task, void *arg) { sched_task_t *t = task; sched_root_task_t *r = NULL; + uintptr_t mask = (uintptr_t) arg; #if SUP_ENABLE == KQ_SUPPORT struct kevent chg[1]; struct timespec timeout = { 0, 0 }; #elif SUP_ENABLE == EP_SUPPORT - struct epoll_event ee; + struct epoll_event ee = { 0 }; int flg = 0; #endif @@ -641,9 +635,11 @@ sched_hook_write(void *task, void *arg __unused) #if SUP_ENABLE == KQ_SUPPORT #ifdef __NetBSD__ - EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_FD(t)); + EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR | mask, + 0, 0, (intptr_t) TASK_FD(t)); #else - EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR, 0, 0, (void*) TASK_FD(t)); + EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD | EV_CLEAR | mask, + 0, 0, (void*) TASK_FD(t)); #endif if (kevent(r->root_kq, chg, 1, NULL, 0, &timeout) == -1) { if (r->root_hooks.hook_exec.exception) @@ -653,8 +649,10 @@ sched_hook_write(void *task, void *arg __unused) return (void*) -1; } #elif SUP_ENABLE == EP_SUPPORT + if (!mask) + mask = EPOLLOUT; ee.data.fd = TASK_FD(t); - ee.events = EPOLLOUT; + ee.events = mask; if (FD_ISSET(TASK_FD(t), &r->root_fds[2])) { flg |= 4; @@ -666,6 +664,7 @@ sched_hook_write(void *task, void *arg __unused) } if (FD_ISSET(TASK_FD(t), &r->root_fds[1])) { flg |= 2; + ee.events |= EPOLLOUT; } if (epoll_ctl(r->root_kq, flg ? EPOLL_CTL_MOD : EPOLL_CTL_ADD, TASK_FD(t), &ee) == -1) { @@ -675,9 +674,14 @@ sched_hook_write(void *task, void *arg __unused) LOGERR; return (void*) -1; } else - FD_SET(TASK_FD(t), &r->root_fds[1]); + if (mask & EPOLLOUT) + FD_SET(TASK_FD(t), &r->root_fds[1]); #else - FD_SET(TASK_FD(t), &r->root_fds[1]); + if (!mask) + FD_SET(TASK_FD(t), &r->root_fds[1]); + else + if (mask & 1) + FD_SET(TASK_FD(t), &r->root_fds[1]); if (TASK_FD(t) >= r->root_kq) r->root_kq = TASK_FD(t) + 1; @@ -812,17 +816,16 @@ sched_hook_proc(void *task, void *arg __unused) void * sched_hook_signal(void *task, void *arg __unused) { -#if SUP_ENABLE == KQ_SUPPORT sched_task_t *t = task; +#if SUP_ENABLE == KQ_SUPPORT struct kevent chg[1]; struct timespec timeout = { 0, 0 }; +#endif if (!t || !TASK_ROOT(t)) return (void*) -1; - /* ignore signal */ - signal(TASK_VAL(t), SIG_IGN); - +#if SUP_ENABLE == KQ_SUPPORT #ifdef __NetBSD__ EV_SET(&chg[0], TASK_VAL(t), EVFILT_SIGNAL, EV_ADD | EV_CLEAR, 0, 0, (intptr_t) TASK_VAL(t)); #else @@ -836,6 +839,11 @@ sched_hook_signal(void *task, void *arg __unused) return (void*) -1; } #endif + + pthread_mutex_lock(&TASK_ROOT(t)->root_sigmtx); + sigaddset(&TASK_ROOT(t)->root_sigset, TASK_VAL(t)); + pthread_mutex_unlock(&TASK_ROOT(t)->root_sigmtx); + return NULL; } @@ -1207,6 +1215,7 @@ fetch_hook_epoll_proceed(int en, struct epoll_event *r } } + ops = EPOLL_CTL_DEL; if (rflg > 1 || wflg > 1) ops = EPOLL_CTL_MOD; @@ -1333,7 +1342,7 @@ sched_hook_fetch(void *root, void *arg __unused) { sched_root_task_t *r = root; sched_task_t *task, *tmp; - struct timespec now, m, mtmp; + struct timespec now, m, mtmp, *tsmin; #if SUP_ENABLE == KQ_SUPPORT struct kevent res[KQ_EVENTS]; struct timespec *timeout; @@ -1376,7 +1385,11 @@ sched_hook_fetch(void *root, void *arg __unused) if (TAILQ_FIRST(&r->root_timer)) { m = r->root_wait; sched_timespecsub(&m, &now, &mtmp); - r->root_wait = mtmp; + if (mtmp.tv_sec < 0 || mtmp.tv_nsec < 0) + /* don't wait for events. we have ready timer */ + sched_timespecclear(&r->root_wait); + else + r->root_wait = mtmp; } else { /* set wait INFTIM */ sched_timespecinf(&r->root_wait); @@ -1387,7 +1400,11 @@ sched_hook_fetch(void *root, void *arg __unused) m = TASK_TS(task); sched_timespecsub(&m, &now, &mtmp); - r->root_wait = mtmp; + if (mtmp.tv_sec < 0 || mtmp.tv_nsec < 0) + /* don't wait for events. we have ready timer */ + sched_timespecclear(&r->root_wait); + else + r->root_wait = mtmp; } else { /* set wait INFTIM */ sched_timespecinf(&r->root_wait); @@ -1396,13 +1413,14 @@ sched_hook_fetch(void *root, void *arg __unused) } else /* no waiting for event, because we have ready task */ sched_timespecclear(&r->root_wait); - if (r->root_wait.tv_sec != -1 && r->root_wait.tv_nsec != -1) { + if (!sched_timespecisinf(&r->root_wait)) { + tsmin = sched_timespecmin(&r->root_wait, &r->root_poll); #if SUP_ENABLE == KQ_SUPPORT - timeout = &r->root_wait; + timeout = tsmin; #elif SUP_ENABLE == EP_SUPPORT - timeout = r->root_wait.tv_sec * 1000 + r->root_wait.tv_nsec / 1000000; + timeout = tsmin->tv_sec * 1000 + tsmin->tv_nsec / 1000000; #else - sched_timespec2val(&r->root_wait, &tv); + sched_timespec2val(tsmin, &tv); timeout = &tv; #endif /* KQ_SUPPORT */ } else if (sched_timespecisinf(&r->root_poll))