version 1.1, 2011/08/05 15:52:00
|
version 1.2, 2011/10/04 12:34:33
|
Line 119 sched_hook_cancel(void *task, void *arg __unused)
|
Line 119 sched_hook_cancel(void *task, void *arg __unused)
|
struct sched_IO *io; |
struct sched_IO *io; |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout; | struct timespec timeout = { 0, 0 }; |
|
|
if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) |
if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) |
return (void*) -1; |
return (void*) -1; |
else |
else |
io = ROOT_DATA(t->task_root); |
io = ROOT_DATA(t->task_root); |
|
|
timespecclear(&timeout); |
|
switch (t->task_type) { |
switch (t->task_type) { |
case taskREAD: |
case taskREAD: |
if (FD_ISSET(TASK_FD(t), &io->wfd)) | #ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, &TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
else | #else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, &TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
| #endif |
kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout); |
kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout); |
|
|
FD_CLR(TASK_FD(t), &io->rfd); |
FD_CLR(TASK_FD(t), &io->rfd); |
break; |
break; |
case taskWRITE: |
case taskWRITE: |
if (FD_ISSET(TASK_FD(t), &io->rfd)) | #ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, &TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (intptr_t) TASK_FD(t)); |
else | #else |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_DELETE, 0, 0, &TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_DELETE, 0, 0, (void*) TASK_FD(t)); |
| #endif |
kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout); |
kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout); |
|
|
FD_CLR(TASK_FD(t), &io->wfd); |
FD_CLR(TASK_FD(t), &io->wfd); |
break; |
break; |
default: |
default: |
Line 165 sched_hook_read(void *task, void *arg __unused)
|
Line 164 sched_hook_read(void *task, void *arg __unused)
|
struct sched_IO *io; |
struct sched_IO *io; |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout; | struct timespec timeout = { 0, 0 }; |
|
|
if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) |
if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) |
return (void*) -1; |
return (void*) -1; |
Line 177 sched_hook_read(void *task, void *arg __unused)
|
Line 176 sched_hook_read(void *task, void *arg __unused)
|
else |
else |
FD_SET(TASK_FD(t), &io->rfd); |
FD_SET(TASK_FD(t), &io->rfd); |
|
|
timespecclear(&timeout); | #ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, &TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (intptr_t) TASK_FD(t)); |
| #else |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_READ, EV_ADD, 0, 0, (void*) TASK_FD(t)); |
| #endif |
if (kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
LOGERR; |
LOGERR; |
return (void*) -1; |
return (void*) -1; |
Line 199 sched_hook_write(void *task, void *arg __unused)
|
Line 201 sched_hook_write(void *task, void *arg __unused)
|
struct sched_IO *io; |
struct sched_IO *io; |
sched_task_t *t = task; |
sched_task_t *t = task; |
struct kevent chg[1]; |
struct kevent chg[1]; |
struct timespec timeout; | struct timespec timeout = { 0, 0 }; |
|
|
if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) |
if (!t || !t->task_root || !ROOT_DATA(t->task_root) || !ROOT_DATLEN(t->task_root)) |
return (void*) -1; |
return (void*) -1; |
Line 211 sched_hook_write(void *task, void *arg __unused)
|
Line 213 sched_hook_write(void *task, void *arg __unused)
|
else |
else |
FD_SET(TASK_FD(t), &io->wfd); |
FD_SET(TASK_FD(t), &io->wfd); |
|
|
timespecclear(&timeout); | #ifdef __NetBSD__ |
EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, &TASK_FD(t)); | EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (intptr_t) TASK_FD(t)); |
| #else |
| EV_SET(&chg[0], TASK_FD(t), EVFILT_WRITE, EV_ADD, 0, 0, (void*) TASK_FD(t)); |
| #endif |
if (kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
if (kevent(t->task_root->root_kq, chg, 1, NULL, 0, &timeout) == -1) { |
LOGERR; |
LOGERR; |
return (void*) -1; |
return (void*) -1; |
Line 304 retry:
|
Line 309 retry:
|
timeout = NULL; |
timeout = NULL; |
if ((en = kevent(r->root_kq, NULL, 0, res, KQ_EVENTS, timeout)) == -1) { |
if ((en = kevent(r->root_kq, NULL, 0, res, KQ_EVENTS, timeout)) == -1) { |
LOGERR; |
LOGERR; |
|
#ifdef NDEBUG |
|
/* kevent no exit by error, if non-debug version */ |
goto retry; |
goto retry; |
|
#else |
|
/* diagnostic exit from scheduler if kevent error occur */ |
|
return NULL; |
|
#endif |
} |
} |
|
|
timespecclear(&nw); | nw.tv_sec = nw.tv_nsec = 0; |
/* Go and catch the cat into pipes ... */ |
/* Go and catch the cat into pipes ... */ |
for (i = 0; i < en; i++) { |
for (i = 0; i < en; i++) { |
memcpy(evt, &res[i], sizeof evt); |
memcpy(evt, &res[i], sizeof evt); |
Line 316 retry:
|
Line 327 retry:
|
switch (res[i].filter) { |
switch (res[i].filter) { |
case EVFILT_READ: |
case EVFILT_READ: |
TAILQ_FOREACH(task, &r->root_read, task_node) { |
TAILQ_FOREACH(task, &r->root_read, task_node) { |
if (TASK_FD(task) != *((int*) res[i].udata)) | if (TASK_FD(task) != ((int) res[i].udata)) |
continue; |
continue; |
/* remove read handle */ |
/* remove read handle */ |
io = ROOT_DATA(task->task_root); |
io = ROOT_DATA(task->task_root); |
FD_CLR(TASK_FD(task), &io->rfd); |
FD_CLR(TASK_FD(task), &io->rfd); |
|
|
TAILQ_REMOVE(&r->root_read, task, task_node); |
TAILQ_REMOVE(&r->root_read, task, task_node); |
task->task_type = taskREADY; | /* if (res[i].flags & EV_EOF) { |
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | task->task_type = taskUNUSE; |
| TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); |
| } else { */ |
| task->task_type = taskREADY; |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
| /* } */ |
break; |
break; |
} |
} |
break; |
break; |
case EVFILT_WRITE: |
case EVFILT_WRITE: |
TAILQ_FOREACH(task, &r->root_write, task_node) { |
TAILQ_FOREACH(task, &r->root_write, task_node) { |
if (TASK_FD(task) != *((int*) res[i].udata)) | if (TASK_FD(task) != ((int) res[i].udata)) |
continue; |
continue; |
/* remove write handle */ |
/* remove write handle */ |
io = ROOT_DATA(task->task_root); |
io = ROOT_DATA(task->task_root); |
FD_CLR(TASK_FD(task), &io->wfd); |
FD_CLR(TASK_FD(task), &io->wfd); |
|
|
TAILQ_REMOVE(&r->root_write, task, task_node); |
TAILQ_REMOVE(&r->root_write, task, task_node); |
task->task_type = taskREADY; | /* if (res[i].flags & EV_EOF) { |
TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); | task->task_type = taskUNUSE; |
| TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); |
| } else { */ |
| task->task_type = taskREADY; |
| TAILQ_INSERT_TAIL(&r->root_ready, task, task_node); |
| /* } */ |
break; |
break; |
} |
} |
break; |
break; |
} |
} |
|
|
if (kevent(r->root_kq, evt, 1, NULL, 0, &nw) == -1) |
if (kevent(r->root_kq, evt, 1, NULL, 0, &nw) == -1) |
LOGERR; |
LOGERR; |
} |
} |
|
|
/* timer update */ | /* timer update & put in ready queue */ |
clock_gettime(CLOCK_MONOTONIC, &nw); |
clock_gettime(CLOCK_MONOTONIC, &nw); |
now.tv_sec = nw.tv_sec; |
now.tv_sec = nw.tv_sec; |
now.tv_usec = nw.tv_nsec / 1000; |
now.tv_usec = nw.tv_nsec / 1000; |
Line 382 retry:
|
Line 402 retry:
|
TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); |
TAILQ_INSERT_TAIL(&r->root_unuse, task, task_node); |
return task; |
return task; |
} |
} |
|
|