+void signal_add(signal_t *sig, signal_cb_t cb, void *data, int signum) {
+ if(sig->cb) {
+ return;
+ }
+
+ sig->cb = cb;
+ sig->data = data;
+ sig->signum = signum;
+ sig->node.data = sig;
+
+ if(pipefd[0] == -1) {
+ pipe_init();
+ }
+
+ signal(sig->signum, signal_handler);
+
+ if(!splay_insert_node(&signal_tree, &sig->node)) {
+ abort();
+ }
+}
+
+void signal_del(signal_t *sig) {
+ if(!sig->cb) {
+ return;
+ }
+
+ signal(sig->signum, SIG_DFL);
+
+ splay_unlink_node(&signal_tree, &sig->node);
+ sig->cb = NULL;
+}
+#endif
+
+static struct timeval *get_time_remaining(struct timeval *diff) {
+ gettimeofday(&now, NULL);
+ struct timeval *tv = NULL;
+
+ while(timeout_tree.head) {
+ timeout_t *timeout = timeout_tree.head->data;
+ timersub(&timeout->tv, &now, diff);
+
+ if(diff->tv_sec < 0) {
+ timeout->cb(timeout->data);
+
+ if(timercmp(&timeout->tv, &now, <)) {
+ timeout_del(timeout);
+ }
+ } else {
+ tv = diff;
+ break;
+ }
+ }
+
+ return tv;
+}
+
+bool event_loop(void) {
+ running = true;
+
+#ifndef HAVE_MINGW
+ fd_set readable;
+ fd_set writable;
+
+ while(running) {
+ struct timeval diff;
+ struct timeval *tv = get_time_remaining(&diff);
+ memcpy(&readable, &readfds, sizeof(readable));
+ memcpy(&writable, &writefds, sizeof(writable));
+
+ int fds = 0;
+
+ if(io_tree.tail) {
+ io_t *last = io_tree.tail->data;
+ fds = last->fd + 1;
+ }
+
+ int n = select(fds, &readable, &writable, NULL, tv);
+
+ if(n < 0) {
+ if(sockwouldblock(sockerrno)) {
+ continue;
+ } else {
+ return false;
+ }
+ }
+
+ if(!n) {
+ continue;
+ }
+
+ for splay_each(io_t, io, &io_tree) {
+ if(FD_ISSET(io->fd, &writable)) {
+ io->cb(io->data, IO_WRITE);
+ } else if(FD_ISSET(io->fd, &readable)) {
+ io->cb(io->data, IO_READ);
+ } else {
+ continue;
+ }
+
+ /*
+ There are scenarios in which the callback will remove another io_t from the tree
+ (e.g. closing a double connection). Since splay_each does not support that, we
+ need to exit the loop now. That's okay, since any remaining events will get picked
+ up by the next select() call.
+ */
+ break;
+ }
+ }
+
+#else
+
+ while(running) {
+ struct timeval diff;
+ struct timeval *tv = get_time_remaining(&diff);
+ DWORD timeout_ms = tv ? (tv->tv_sec * 1000 + tv->tv_usec / 1000 + 1) : WSA_INFINITE;
+
+ if(!event_count) {
+ Sleep(timeout_ms);
+ continue;
+ }
+
+ /*
+ For some reason, Microsoft decided to make the FD_WRITE event edge-triggered instead of level-triggered,
+ which is the opposite of what select() does. In practice, that means that if a FD_WRITE event triggers,
+ it will never trigger again until a send() returns EWOULDBLOCK. Since the semantics of this event loop
+ is that write events are level-triggered (i.e. they continue firing until the socket is full), we need
+ to emulate these semantics by making sure we fire each IO_WRITE that is still writeable.
+
+ Note that technically FD_CLOSE has the same problem, but it's okay because user code does not rely on
+ this event being fired again if ignored.
+ */
+ io_t *writeable_io = NULL;
+
+ for splay_each(io_t, io, &io_tree)
+ if(io->flags & IO_WRITE && send(io->fd, NULL, 0, 0) == 0) {
+ writeable_io = io;
+ break;
+ }
+
+ if(writeable_io) {
+ writeable_io->cb(writeable_io->data, IO_WRITE);
+ continue;
+ }
+
+ WSAEVENT *events = xmalloc(event_count * sizeof(*events));
+ DWORD event_index = 0;
+
+ for splay_each(io_t, io, &io_tree) {
+ events[event_index] = io->event;
+ event_index++;
+ }
+
+ DWORD result = WSAWaitForMultipleEvents(event_count, events, FALSE, timeout_ms, FALSE);
+
+ WSAEVENT event;
+
+ if(result >= WSA_WAIT_EVENT_0 && result < WSA_WAIT_EVENT_0 + event_count) {
+ event = events[result - WSA_WAIT_EVENT_0];
+ }
+
+ free(events);
+
+ if(result == WSA_WAIT_TIMEOUT) {
+ continue;
+ }
+
+ if(result < WSA_WAIT_EVENT_0 || result >= WSA_WAIT_EVENT_0 + event_count) {
+ return false;
+ }
+
+ io_t *io = splay_search(&io_tree, &((io_t) {
+ .event = event
+ }));
+
+ if(!io) {
+ abort();
+ }
+
+ if(io->fd == -1) {
+ io->cb(io->data, 0);
+ } else {
+ WSANETWORKEVENTS network_events;
+
+ if(WSAEnumNetworkEvents(io->fd, io->event, &network_events) != 0) {
+ return false;
+ }
+
+ if(network_events.lNetworkEvents & READ_EVENTS) {
+ io->cb(io->data, IO_READ);
+ }
+
+ /*
+ The fd might be available for write too. However, if we already fired the read callback, that
+ callback might have deleted the io (e.g. through terminate_connection()), so we can't fire the
+ write callback here. Instead, we loop back and let the writable io loop above handle it.
+ */