/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/ /*** This file is part of systemd. Copyright 2013 Lennart Poettering systemd is free software; you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation; either version 2.1 of the License, or (at your option) any later version. systemd is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more details. You should have received a copy of the GNU Lesser General Public License along with systemd; If not, see . ***/ #include #include #include #include "macro.h" #include "refcnt.h" #include "prioq.h" #include "hashmap.h" #include "util.h" #include "time-util.h" #include "sd-event.h" #define EPOLL_QUEUE_MAX 64 typedef enum EventSourceType { SOURCE_IO, SOURCE_MONOTONIC, SOURCE_REALTIME, SOURCE_SIGNAL, SOURCE_CHILD, SOURCE_DEFER } EventSourceType; struct sd_event_source { RefCount n_ref; sd_event *event; void *userdata; sd_prepare_handler_t prepare; EventSourceType type:4; sd_event_mute_t mute:3; bool pending:1; int priority; unsigned pending_index; unsigned prepare_index; unsigned pending_iteration; unsigned prepare_iteration; union { struct { sd_io_handler_t callback; int fd; uint32_t events; uint32_t revents; bool registered:1; } io; struct { sd_time_handler_t callback; usec_t next; unsigned prioq_index; } time; struct { sd_signal_handler_t callback; struct signalfd_siginfo siginfo; int sig; } signal; struct { sd_child_handler_t callback; siginfo_t siginfo; pid_t pid; int options; } child; struct { sd_defer_handler_t callback; } defer; }; }; struct sd_event { RefCount n_ref; int epoll_fd; int signal_fd; int realtime_fd; int monotonic_fd; Prioq *pending; Prioq *prepare; Prioq *monotonic; Prioq *realtime; sigset_t sigset; sd_event_source **signal_sources; Hashmap *child_sources; unsigned n_unmuted_child_sources; unsigned iteration; unsigned processed_children; usec_t realtime_next, monotonic_next; bool quit; }; static int pending_prioq_compare(const void *a, const void *b) { const sd_event_source *x = a, *y = b; assert(x->pending); assert(y->pending); /* Unmuted ones first */ if (x->mute != SD_EVENT_MUTED && y->mute == SD_EVENT_MUTED) return -1; if (x->mute == SD_EVENT_MUTED && y->mute != SD_EVENT_MUTED) return 1; /* Lower priority values first */ if (x->priority < y->priority) return -1; if (x->priority > y->priority) return 1; /* Older entries first */ if (x->pending_iteration < y->pending_iteration) return -1; if (x->pending_iteration > y->pending_iteration) return 1; /* Stability for the rest */ if (x < y) return -1; if (y > x) return 1; return 0; } static int prepare_prioq_compare(const void *a, const void *b) { const sd_event_source *x = a, *y = b; assert(x->prepare); assert(y->prepare); /* Move most recently prepared ones last, so that we can stop * preparing as soon as we hit one that has already been * prepared in the current iteration */ if (x->prepare_iteration < y->prepare_iteration) return -1; if (x->prepare_iteration > y->prepare_iteration) return 1; /* Unmuted ones first */ if (x->mute != SD_EVENT_MUTED && y->mute == SD_EVENT_MUTED) return -1; if (x->mute == SD_EVENT_MUTED && y->mute != SD_EVENT_MUTED) return 1; /* Lower priority values first */ if (x->priority < y->priority) return -1; if (x->priority > y->priority) return 1; /* Stability for the rest */ if (x < y) return -1; if (y > x) return 1; return 0; } static int time_prioq_compare(const void *a, const void *b) { const sd_event_source *x = a, *y = b; assert(x->type == SOURCE_MONOTONIC || x->type == SOURCE_REALTIME); assert(y->type == SOURCE_MONOTONIC || y->type == SOURCE_REALTIME); /* Unmuted ones first */ if (x->mute != SD_EVENT_MUTED && y->mute == SD_EVENT_MUTED) return -1; if (x->mute == SD_EVENT_MUTED && y->mute != SD_EVENT_MUTED) return 1; /* Move the pending ones to the end */ if (!x->pending && y->pending) return -1; if (x->pending && !y->pending) return 1; /* Order by time */ if (x->time.next < y->time.next) return -1; if (x->time.next > y->time.next) return -1; /* Stability for the rest */ if (x < y) return -1; if (y > x) return 1; return 0; } static void event_free(sd_event *e) { assert(e); if (e->epoll_fd >= 0) close_nointr_nofail(e->epoll_fd); if (e->signal_fd >= 0) close_nointr_nofail(e->signal_fd); if (e->realtime_fd >= 0) close_nointr_nofail(e->realtime_fd); if (e->monotonic_fd >= 0) close_nointr_nofail(e->monotonic_fd); prioq_free(e->pending); prioq_free(e->prepare); prioq_free(e->monotonic); prioq_free(e->realtime); free(e->signal_sources); hashmap_free(e->child_sources); free(e); } int sd_event_new(sd_event** ret) { sd_event *e; int r; if (!ret) return -EINVAL; e = new0(sd_event, 1); if (!e) return -ENOMEM; e->n_ref = REFCNT_INIT; e->signal_fd = e->realtime_fd = e->monotonic_fd = e->epoll_fd = -1; e->realtime_next = e->monotonic_next = (usec_t) -1; assert_se(sigemptyset(&e->sigset) == 0); e->pending = prioq_new(pending_prioq_compare); if (!e->pending) { r = -ENOMEM; goto fail; } e->epoll_fd = epoll_create1(EPOLL_CLOEXEC); if (e->epoll_fd < 0) { r = -errno; goto fail; } *ret = e; return 0; fail: event_free(e); return r; } sd_event* sd_event_ref(sd_event *e) { if (!e) return NULL; assert_se(REFCNT_INC(e->n_ref) >= 2); return e; } sd_event* sd_event_unref(sd_event *e) { if (!e) return NULL; if (REFCNT_DEC(e->n_ref) <= 0) event_free(e); return NULL; } static int source_io_unregister(sd_event_source *s) { int r; assert(s); assert(s->type == SOURCE_IO); if (!s->io.registered) return 0; r = epoll_ctl(s->event->epoll_fd, EPOLL_CTL_DEL, s->io.fd, NULL); if (r < 0) return -errno; s->io.registered = false; return 0; } static int source_io_register(sd_event_source *s, sd_event_mute_t m, uint32_t events) { struct epoll_event ev = {}; int r; assert(s); assert(s->type == SOURCE_IO); assert(m != SD_EVENT_MUTED); ev.events = events; ev.data.ptr = s; if (m == SD_EVENT_ONESHOT) ev.events |= EPOLLONESHOT; if (s->io.registered) r = epoll_ctl(s->event->epoll_fd, EPOLL_CTL_MOD, s->io.fd, &ev); else r = epoll_ctl(s->event->epoll_fd, EPOLL_CTL_ADD, s->io.fd, &ev); if (r < 0) return -errno; s->io.registered = true; return 0; } static void source_free(sd_event_source *s) { assert(s); if (s->event) { switch (s->type) { case SOURCE_IO: if (s->io.fd >= 0) source_io_unregister(s); break; case SOURCE_MONOTONIC: prioq_remove(s->event->monotonic, s, &s->time.prioq_index); break; case SOURCE_REALTIME: prioq_remove(s->event->realtime, s, &s->time.prioq_index); break; case SOURCE_SIGNAL: if (s->signal.sig > 0) { if (s->signal.sig != SIGCHLD || s->event->n_unmuted_child_sources == 0) assert_se(sigdelset(&s->event->sigset, s->signal.sig) == 0); if (s->event->signal_sources) s->event->signal_sources[s->signal.sig] = NULL; } break; case SOURCE_CHILD: if (s->child.pid > 0) { if (s->mute != SD_EVENT_MUTED) { assert(s->event->n_unmuted_child_sources > 0); s->event->n_unmuted_child_sources--; } if (!s->event->signal_sources || !s->event->signal_sources[SIGCHLD]) assert_se(sigdelset(&s->event->sigset, SIGCHLD) == 0); hashmap_remove(s->event->child_sources, INT_TO_PTR(s->child.pid)); } break; } if (s->pending) prioq_remove(s->event->pending, s, &s->pending_index); if (s->prepare) prioq_remove(s->event->prepare, s, &s->prepare_index); sd_event_unref(s->event); } free(s); } static int source_set_pending(sd_event_source *s, bool b) { int r; assert(s); if (s->pending == b) return 0; s->pending = b; if (b) { s->pending_iteration = s->event->iteration; r = prioq_put(s->event->pending, s, &s->pending_index); if (r < 0) { s->pending = false; return r; } } else assert_se(prioq_remove(s->event->pending, s, &s->pending_index)); return 0; } static sd_event_source *source_new(sd_event *e, EventSourceType type) { sd_event_source *s; assert(e); s = new0(sd_event_source, 1); if (!s) return NULL; s->n_ref = REFCNT_INIT; s->event = sd_event_ref(e); s->type = type; s->mute = SD_EVENT_UNMUTED; s->pending_index = s->prepare_index = PRIOQ_IDX_NULL; return s; } int sd_event_add_io( sd_event *e, int fd, uint32_t events, sd_io_handler_t callback, void *userdata, sd_event_source **ret) { sd_event_source *s; int r; if (!e) return -EINVAL; if (fd < 0) return -EINVAL; if (events & ~(EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLPRI|EPOLLERR|EPOLLHUP)) return -EINVAL; if (!callback) return -EINVAL; if (!ret) return -EINVAL; s = source_new(e, SOURCE_IO); if (!s) return -ENOMEM; s->io.fd = fd; s->io.events = events; s->io.callback = callback; s->userdata = userdata; r = source_io_register(s, s->mute, events); if (r < 0) { source_free(s); return -errno; } *ret = s; return 0; } static int event_setup_timer_fd( sd_event *e, EventSourceType type, int *timer_fd, clockid_t id) { struct epoll_event ev = {}; int r, fd; assert(e); assert(timer_fd); if (_likely_(*timer_fd >= 0)) return 0; fd = timerfd_create(id, TFD_NONBLOCK|TFD_CLOEXEC); if (fd < 0) return -errno; ev.events = EPOLLIN; ev.data.ptr = INT_TO_PTR(type); r = epoll_ctl(e->epoll_fd, EPOLL_CTL_ADD, fd, &ev); if (r < 0) { close_nointr_nofail(fd); return -errno; } *timer_fd = fd; return 0; } static int event_add_time_internal( sd_event *e, EventSourceType type, int *timer_fd, clockid_t id, Prioq **prioq, uint64_t usec, sd_time_handler_t callback, void *userdata, sd_event_source **ret) { sd_event_source *s; int r; if (!e) return -EINVAL; if (!callback) return -EINVAL; if (!ret) return -EINVAL; assert(timer_fd); assert(prioq); if (!*prioq) { *prioq = prioq_new(time_prioq_compare); if (!*prioq) return -ENOMEM; } if (*timer_fd < 0) { r = event_setup_timer_fd(e, type, timer_fd, id); if (r < 0) return r; } s = source_new(e, type); if (!s) return -ENOMEM; s->time.next = usec; s->time.callback = callback; s->time.prioq_index = PRIOQ_IDX_NULL; s->userdata = userdata; r = prioq_put(*prioq, s, &s->time.prioq_index); if (r < 0) { source_free(s); return r; } *ret = s; return 0; } int sd_event_add_monotonic(sd_event *e, uint64_t usec, sd_time_handler_t callback, void *userdata, sd_event_source **ret) { return event_add_time_internal(e, SOURCE_MONOTONIC, &e->monotonic_fd, CLOCK_MONOTONIC, &e->monotonic, usec, callback, userdata, ret); } int sd_event_add_realtime(sd_event *e, uint64_t usec, sd_time_handler_t callback, void *userdata, sd_event_source **ret) { return event_add_time_internal(e, SOURCE_REALTIME, &e->realtime_fd, CLOCK_REALTIME, &e->realtime, usec, callback, userdata, ret); } static int event_update_signal_fd(sd_event *e) { struct epoll_event ev = {}; bool add_to_epoll; int r; assert(e); add_to_epoll = e->signal_fd < 0; r = signalfd(e->signal_fd, &e->sigset, SFD_NONBLOCK|SFD_CLOEXEC); if (r < 0) return -errno; e->signal_fd = r; if (!add_to_epoll) return 0; ev.events = EPOLLIN; ev.data.ptr = INT_TO_PTR(SOURCE_SIGNAL); r = epoll_ctl(e->epoll_fd, EPOLL_CTL_ADD, e->signal_fd, &ev); if (r < 0) { close_nointr_nofail(e->signal_fd); e->signal_fd = -1; return -errno; } return 0; } int sd_event_add_signal(sd_event *e, int sig, sd_signal_handler_t callback, void *userdata, sd_event_source **ret) { sd_event_source *s; int r; if (!e) return -EINVAL; if (sig <= 0) return -EINVAL; if (sig >= _NSIG) return -EINVAL; if (!callback) return -EINVAL; if (!ret) return -EINVAL; if (!e->signal_sources) { e->signal_sources = new0(sd_event_source*, _NSIG); if (!e->signal_sources) return -ENOMEM; } else if (e->signal_sources[sig]) return -EBUSY; s = source_new(e, SOURCE_SIGNAL); if (!s) return -ENOMEM; s->signal.sig = sig; s->signal.callback = callback; s->userdata = userdata; e->signal_sources[sig] = s; assert_se(sigaddset(&e->sigset, sig) == 0); if (sig != SIGCHLD || e->n_unmuted_child_sources == 0) { r = event_update_signal_fd(e); if (r < 0) { source_free(s); return r; } } *ret = s; return 0; } int sd_event_add_child(sd_event *e, pid_t pid, int options, sd_child_handler_t callback, void *userdata, sd_event_source **ret) { sd_event_source *s; int r; if (!e) return -EINVAL; if (pid <= 1) return -EINVAL; if (options & ~(WEXITED|WSTOPPED|WCONTINUED)) return -EINVAL; if (!callback) return -EINVAL; if (!ret) return -EINVAL; r = hashmap_ensure_allocated(&e->child_sources, trivial_hash_func, trivial_compare_func); if (r < 0) return r; if (hashmap_contains(e->child_sources, INT_TO_PTR(pid))) return -EBUSY; s = source_new(e, SOURCE_CHILD); if (!s) return -ENOMEM; s->child.pid = pid; s->child.options = options; s->child.callback = callback; s->userdata = userdata; r = hashmap_put(e->child_sources, INT_TO_PTR(pid), s); if (r < 0) { source_free(s); return r; } e->n_unmuted_child_sources ++; assert_se(sigaddset(&e->sigset, SIGCHLD) == 0); if (!e->signal_sources || !e->signal_sources[SIGCHLD]) { r = event_update_signal_fd(e); if (r < 0) { source_free(s); return -errno; } } *ret = s; return 0; } int sd_event_add_defer(sd_event *e, sd_defer_handler_t callback, void *userdata, sd_event_source **ret) { sd_event_source *s; int r; if (!e) return -EINVAL; if (!ret) return -EINVAL; s = source_new(e, SOURCE_DEFER); if (!s) return -ENOMEM; s->defer.callback = callback; s->userdata = userdata; r = source_set_pending(s, true); if (r < 0) { source_free(s); return r; } *ret = s; return 0; } sd_event_source* sd_event_source_ref(sd_event_source *s) { if (!s) return NULL; assert_se(REFCNT_INC(s->n_ref) >= 2); return s; } sd_event_source* sd_event_source_unref(sd_event_source *s) { if (!s) return NULL; if (REFCNT_DEC(s->n_ref) <= 0) source_free(s); return NULL; } int sd_event_source_get_pending(sd_event_source *s) { if (!s) return -EINVAL; return s->pending; } int sd_event_source_get_io_fd(sd_event_source *s) { if (!s) return -EINVAL; if (s->type != SOURCE_IO) return -EDOM; return s->io.fd; } int sd_event_source_get_io_events(sd_event_source *s, uint32_t* events) { if (!s) return -EINVAL; if (s->type != SOURCE_IO) return -EDOM; if (!events) return -EINVAL; *events = s->io.events; return 0; } int sd_event_source_set_io_events(sd_event_source *s, uint32_t events) { int r; if (!s) return -EINVAL; if (!s->type != SOURCE_IO) return -EDOM; if (events & ~(EPOLLIN|EPOLLOUT|EPOLLRDHUP|EPOLLPRI|EPOLLERR|EPOLLHUP)) return -EINVAL; if (s->io.events == events) return 0; if (s->mute != SD_EVENT_MUTED) { r = source_io_register(s, s->io.events, events); if (r < 0) return r; } s->io.events = events; return 0; } int sd_event_source_get_io_revents(sd_event_source *s, uint32_t* revents) { if (!s) return -EINVAL; if (s->type != SOURCE_IO) return -EDOM; if (!revents) return -EINVAL; if (!s->pending) return -ENODATA; *revents = s->io.revents; return 0; } int sd_event_source_get_signal(sd_event_source *s) { if (!s) return -EINVAL; if (s->type != SOURCE_SIGNAL) return -EDOM; return s->signal.sig; } int sd_event_source_get_priority(sd_event_source *s, int *priority) { if (!s) return -EINVAL; return s->priority; } int sd_event_source_set_priority(sd_event_source *s, int priority) { if (!s) return -EINVAL; if (s->priority == priority) return 0; s->priority = priority; if (s->pending) assert_se(prioq_reshuffle(s->event->pending, s, &s->pending_index) == 0); if (s->prepare) assert_se(prioq_reshuffle(s->event->prepare, s, &s->prepare_index) == 0); return 0; } int sd_event_source_get_mute(sd_event_source *s, sd_event_mute_t *m) { if (!s) return -EINVAL; if (!m) return -EINVAL; *m = s->mute; return 0; } int sd_event_source_set_mute(sd_event_source *s, sd_event_mute_t m) { int r; if (!s) return -EINVAL; if (m != SD_EVENT_MUTED && m != SD_EVENT_UNMUTED && !SD_EVENT_ONESHOT) return -EINVAL; if (s->mute == m) return 0; if (m == SD_EVENT_MUTED) { switch (s->type) { case SOURCE_IO: r = source_io_unregister(s); if (r < 0) return r; s->mute = m; break; case SOURCE_MONOTONIC: s->mute = m; prioq_reshuffle(s->event->monotonic, s, &s->time.prioq_index); break; case SOURCE_REALTIME: s->mute = m; prioq_reshuffle(s->event->realtime, s, &s->time.prioq_index); break; case SOURCE_SIGNAL: s->mute = m; if (s->signal.sig != SIGCHLD || s->event->n_unmuted_child_sources == 0) { assert_se(sigdelset(&s->event->sigset, s->signal.sig) == 0); event_update_signal_fd(s->event); } break; case SOURCE_CHILD: s->mute = m; assert(s->event->n_unmuted_child_sources > 0); s->event->n_unmuted_child_sources--; if (!s->event->signal_sources || !s->event->signal_sources[SIGCHLD]) { assert_se(sigdelset(&s->event->sigset, SIGCHLD) == 0); event_update_signal_fd(s->event); } break; case SOURCE_DEFER: s->mute = m; break; } } else { switch (s->type) { case SOURCE_IO: r = source_io_register(s, m, s->io.events); if (r < 0) return r; s->mute = m; break; case SOURCE_MONOTONIC: s->mute = m; prioq_reshuffle(s->event->monotonic, s, &s->time.prioq_index); break; case SOURCE_REALTIME: s->mute = m; prioq_reshuffle(s->event->realtime, s, &s->time.prioq_index); break; case SOURCE_SIGNAL: s->mute = m; if (s->signal.sig != SIGCHLD || s->event->n_unmuted_child_sources == 0) { assert_se(sigaddset(&s->event->sigset, s->signal.sig) == 0); event_update_signal_fd(s->event); } break; case SOURCE_CHILD: s->mute = m; if (s->mute == SD_EVENT_MUTED) { s->event->n_unmuted_child_sources++; if (!s->event->signal_sources || !s->event->signal_sources[SIGCHLD]) { assert_se(sigaddset(&s->event->sigset, SIGCHLD) == 0); event_update_signal_fd(s->event); } } break; case SOURCE_DEFER: s->mute = m; break; } } if (s->pending) prioq_reshuffle(s->event->pending, s, &s->pending_index); if (s->prepare) prioq_reshuffle(s->event->prepare, s, &s->prepare_index); return 0; } int sd_event_source_get_time(sd_event_source *s, uint64_t *usec) { if (!s) return -EINVAL; if (!usec) return -EINVAL; if (s->type != SOURCE_REALTIME && s->type != SOURCE_MONOTONIC) return -EDOM; *usec = s->time.next; return 0; } int sd_event_source_set_time(sd_event_source *s, uint64_t usec) { if (!s) return -EINVAL; if (s->type != SOURCE_REALTIME && s->type != SOURCE_MONOTONIC) return -EDOM; if (s->time.next == usec) return 0; s->time.next = usec; if (s->type == SOURCE_REALTIME) prioq_reshuffle(s->event->realtime, s, &s->time.prioq_index); else prioq_reshuffle(s->event->monotonic, s, &s->time.prioq_index); return 0; } int sd_event_source_set_prepare(sd_event_source *s, sd_prepare_handler_t callback) { int r; if (!s) return -EINVAL; if (s->prepare == callback) return 0; if (callback && s->prepare) { s->prepare = callback; return 0; } r = prioq_ensure_allocated(&s->event->prepare, prepare_prioq_compare); if (r < 0) return r; s->prepare = callback; if (callback) { r = prioq_put(s->event->prepare, s, &s->prepare_index); if (r < 0) return r; } else prioq_remove(s->event->prepare, s, &s->prepare_index); return 0; } void* sd_event_source_get_userdata(sd_event_source *s) { if (!s) return NULL; return s->userdata; } static int event_arm_timer( sd_event *e, int timer_fd, Prioq *prioq, usec_t *next) { struct itimerspec its = {}; sd_event_source *s; int r; assert_se(e); assert_se(next); s = prioq_peek(prioq); if (!s || s->mute == SD_EVENT_MUTED) return 0; if (*next == s->time.next) return 0; assert_se(timer_fd >= 0); if (s->time.next == 0) { /* We don' want to disarm here, just mean some time looooong ago. */ its.it_value.tv_sec = 0; its.it_value.tv_nsec = 1; } else timespec_store(&its.it_value, s->time.next); r = timerfd_settime(timer_fd, TFD_TIMER_ABSTIME, &its, NULL); if (r < 0) return r; *next = s->time.next; return 0; } static int process_io(sd_event *e, sd_event_source *s, uint32_t events) { assert(e); assert(s); assert(s->type == SOURCE_IO); s->io.revents = events; /* If this is a oneshot event source, then we added it to the epoll with EPOLLONESHOT, hence we know it's not registered anymore. We can save a syscall here... */ if (s->mute == SD_EVENT_ONESHOT) s->io.registered = false; return source_set_pending(s, true); } static int flush_timer(sd_event *e, int fd, uint32_t events) { uint64_t x; ssize_t ss; assert(e); if (events != EPOLLIN) return -EIO; ss = read(fd, &x, sizeof(x)); if (ss < 0) { if (errno == EAGAIN || errno == EINTR) return 0; return -errno; } if (ss != sizeof(x)) return -EIO; return 0; } static int process_timer(sd_event *e, usec_t n, Prioq *prioq) { sd_event_source *s; int r; assert(e); for (;;) { s = prioq_peek(prioq); if (!s || s->time.next > n || s->mute == SD_EVENT_MUTED || s->pending) break; r = source_set_pending(s, true); if (r < 0) return r; r = prioq_reshuffle(prioq, s, &s->time.prioq_index); if (r < 0) return r; } return 0; } static int process_child(sd_event *e) { sd_event_source *s; Iterator i; int r; assert(e); /* So, this is ugly. We iteratively invoke waitid() with P_PID + WNOHANG for each PID we wait for, instead of using P_ALL. This is because we only want to get child information of very specific child processes, and not all of them. We might not have processed the SIGCHLD even of a previous invocation and we don't want to maintain a unbounded *per-child* event queue, hence we really don't want anything flushed out of the kernel's queue that we don't care about. Since this is O(n) this means that if you have a lot of processes you probably want to handle SIGCHLD yourself. */ HASHMAP_FOREACH(s, e->child_sources, i) { assert(s->type == SOURCE_CHILD); if (s->pending) continue; if (s->mute == SD_EVENT_MUTED) continue; zero(s->child.siginfo); r = waitid(P_PID, s->child.pid, &s->child.siginfo, WNOHANG|s->child.options); if (r < 0) return -errno; if (s->child.siginfo.si_pid != 0) { r = source_set_pending(s, true); if (r < 0) return r; } } e->processed_children = e->iteration; return 0; } static int process_signal(sd_event *e, uint32_t events) { struct signalfd_siginfo si; bool read_one = false; ssize_t ss; int r; if (events != EPOLLIN) return -EIO; for (;;) { sd_event_source *s; ss = read(e->signal_fd, &si, sizeof(si)); if (ss < 0) { if (errno == EAGAIN || errno == EINTR) return read_one; return -errno; } if (ss != sizeof(si)) return -EIO; read_one = true; if (si.ssi_signo == SIGCHLD) { r = process_child(e); if (r < 0) return r; if (r > 0 || !e->signal_sources[si.ssi_signo]) continue; } else { s = e->signal_sources[si.ssi_signo]; if (!s) return -EIO; } s->signal.siginfo = si; r = source_set_pending(s, true); if (r < 0) return r; } return 0; } static int source_dispatch(sd_event_source *s) { int r; assert(s); assert(s->pending); r = source_set_pending(s, false); if (r < 0) return r; if (s->mute == SD_EVENT_ONESHOT) { r = sd_event_source_set_mute(s, SD_EVENT_MUTED); if (r < 0) return r; } switch (s->type) { case SOURCE_IO: r = s->io.callback(s, s->io.fd, s->io.revents, s->userdata); break; case SOURCE_MONOTONIC: r = s->time.callback(s, s->time.next, s->userdata); break; case SOURCE_REALTIME: r = s->time.callback(s, s->time.next, s->userdata); break; case SOURCE_SIGNAL: r = s->signal.callback(s, &s->signal.siginfo, s->userdata); break; case SOURCE_CHILD: r = s->child.callback(s, &s->child.siginfo, s->userdata); break; case SOURCE_DEFER: r = s->defer.callback(s, s->userdata); break; } return r; } static int event_prepare(sd_event *e) { int r; assert(e); for (;;) { sd_event_source *s; s = prioq_peek(e->prepare); if (!s || s->prepare_iteration == e->iteration || s->mute == SD_EVENT_MUTED) break; s->prepare_iteration = e->iteration; r = prioq_reshuffle(e->prepare, s, &s->prepare_index); if (r < 0) return r; assert(s->prepare); r = s->prepare(s, s->userdata); if (r < 0) return r; } return 0; } int sd_event_run(sd_event *e, uint64_t timeout) { struct epoll_event ev_queue[EPOLL_QUEUE_MAX]; sd_event_source *p; int r, i, m; dual_timestamp n; if (!e) return -EINVAL; if (e->quit) return -ESTALE; e->iteration++; r = event_prepare(e); if (r < 0) return r; r = event_arm_timer(e, e->monotonic_fd, e->monotonic, &e->monotonic_next); if (r < 0) return r; r = event_arm_timer(e, e->realtime_fd, e->realtime, &e->realtime_next); if (r < 0) return r; if (e->iteration == 1 && !hashmap_isempty(e->child_sources)) /* On the first iteration, there might be already some * zombies for us to care for, hence, don't wait */ timeout = 0; else { p = prioq_peek(e->pending); if (p && p->mute != SD_EVENT_MUTED) timeout = 0; } m = epoll_wait(e->epoll_fd, ev_queue, EPOLL_QUEUE_MAX, timeout == (uint64_t) -1 ? -1 : (int) ((timeout + USEC_PER_MSEC - 1) / USEC_PER_MSEC)); if (m < 0) return m; dual_timestamp_get(&n); for (i = 0; i < m; i++) { if (ev_queue[i].data.ptr == INT_TO_PTR(SOURCE_MONOTONIC)) r = flush_timer(e, e->monotonic_fd, ev_queue[i].events); else if (ev_queue[i].data.ptr == INT_TO_PTR(SOURCE_REALTIME)) r = flush_timer(e, e->realtime_fd, ev_queue[i].events); else if (ev_queue[i].data.ptr == INT_TO_PTR(SOURCE_SIGNAL)) r = process_signal(e, ev_queue[i].events); else r = process_io(e, ev_queue[i].data.ptr, ev_queue[i].events); if (r < 0) return r; } r = process_timer(e, n.monotonic, e->monotonic); if (r < 0) return r; r = process_timer(e, n.realtime, e->realtime); if (r < 0) return r; if (e->iteration == 1 && e->processed_children != 1) { /* On the first iteration, make sure we really process * all children which might already be zombies. */ r = process_child(e); if (r < 0) return r; } p = prioq_peek(e->pending); if (!p || p->mute == SD_EVENT_MUTED) return 0; return source_dispatch(p); } int sd_event_loop(sd_event *e) { int r; if (!e) return -EINVAL; while (!e->quit) { r = sd_event_run(e, (uint64_t) -1); if (r < 0) return r; } return 0; } int sd_event_quit(sd_event *e) { if (!e) return EINVAL; return e->quit; } int sd_event_request_quit(sd_event *e) { if (!e) return -EINVAL; e->quit = true; return 0; } sd_event *sd_event_get(sd_event_source *s) { if (!s) return NULL; return s->event; }