X-Git-Url: https://www.tinc-vpn.org/git/browse?a=blobdiff_plain;f=src%2Fevent.c;h=1283edd332b8334fa8fc78a6faedb03d2a397516;hb=refs%2Fheads%2F1.1;hp=c547c46d1f2f5d5a5077d83c335d823c634bb601;hpb=b322126576c52ffbc9a91d818125fc461fa17d1b;p=tinc diff --git a/src/event.c b/src/event.c index c547c46d..05ddb47f 100644 --- a/src/event.c +++ b/src/event.c @@ -1,6 +1,6 @@ /* - event.c -- I/O, timeout and signal event handling - Copyright (C) 2012-2021 Guus Sliepen + event.c -- I/O, timeout, and event handling + Copyright (C) 2012-2022 Guus Sliepen This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by @@ -18,46 +18,13 @@ */ #include "system.h" -#include "dropin.h" - -#ifdef HAVE_SYS_EPOLL_H -#include -#endif #include "event.h" -#include "utils.h" -#include "net.h" struct timeval now; -#ifndef HAVE_MINGW - -#ifdef HAVE_SYS_EPOLL_H -static int epollset = 0; -#else -static fd_set readfds; -static fd_set writefds; -#endif - -#else -static const long READ_EVENTS = FD_READ | FD_ACCEPT | FD_CLOSE; -static const long WRITE_EVENTS = FD_WRITE | FD_CONNECT; -static DWORD event_count = 0; -#endif -static bool running; - -#ifdef HAVE_SYS_EPOLL_H -static inline int event_epoll_init() { - /* NOTE: 1024 limit is only used on ancient (pre 2.6.27) kernels. - Decent kernels will ignore this value making it unlimited. - epoll_create1 might be better, but these kernels would not be supported - in that case. - */ - return epoll_create(1024); -} -#endif static int io_compare(const io_t *a, const io_t *b) { -#ifndef HAVE_MINGW +#ifndef HAVE_WINDOWS return a->fd - b->fd; #else @@ -93,158 +60,24 @@ static int timeout_compare(const timeout_t *a, const timeout_t *b) { return 1; } - if(a < b) { + uintptr_t ap = (uintptr_t)a; + uintptr_t bp = (uintptr_t)b; + + if(ap < bp) { return -1; } - if(a > b) { + if(ap > bp) { return 1; } return 0; } -static splay_tree_t io_tree = {.compare = (splay_compare_t)io_compare}; -static splay_tree_t timeout_tree = {.compare = (splay_compare_t)timeout_compare}; - -void io_add(io_t *io, io_cb_t cb, void *data, int fd, int flags) { - if(io->cb) { - return; - } - - io->fd = fd; -#ifdef HAVE_MINGW - - if(io->fd != -1) { - io->event = WSACreateEvent(); - - if(io->event == WSA_INVALID_EVENT) { - abort(); - } - } - - event_count++; -#endif - io->cb = cb; - io->data = data; - io->node.data = io; - - io_set(io, flags); - -#ifndef HAVE_SYS_EPOLL_H - - if(!splay_insert_node(&io_tree, &io->node)) { - abort(); - } - -#endif -} - -#ifdef HAVE_MINGW -void io_add_event(io_t *io, io_cb_t cb, void *data, WSAEVENT event) { - io->event = event; - io_add(io, cb, data, -1, 0); -} -#endif - -void io_set(io_t *io, int flags) { -#ifdef HAVE_SYS_EPOLL_H - - if(!epollset) { - epollset = event_epoll_init(); - } - -#endif - - if(flags == io->flags) { - return; - } - - io->flags = flags; - - if(io->fd == -1) { - return; - } - -#ifndef HAVE_MINGW -#ifdef HAVE_SYS_EPOLL_H - epoll_ctl(epollset, EPOLL_CTL_DEL, io->fd, NULL); - - struct epoll_event ev = { - .events = 0, - .data.ptr = io, - }; - - if(flags & IO_READ) { - ev.events |= EPOLLIN; - } - - if(flags & IO_WRITE) { - ev.events |= EPOLLOUT; - } else if(ev.events == 0) { - io_tree.generation++; - return; - } - - if(epoll_ctl(epollset, EPOLL_CTL_ADD, io->fd, &ev) < 0) { - perror("epoll_ctl_add"); - } - -#else - - if(flags & IO_READ) { - FD_SET(io->fd, &readfds); - } else { - FD_CLR(io->fd, &readfds); - } - - if(flags & IO_WRITE) { - FD_SET(io->fd, &writefds); - } else { - FD_CLR(io->fd, &writefds); - } - -#endif -#else - long events = 0; - - if(flags & IO_WRITE) { - events |= WRITE_EVENTS; - } - - if(flags & IO_READ) { - events |= READ_EVENTS; - } - - if(WSAEventSelect(io->fd, io->event, events) != 0) { - abort(); - } - -#endif -} - -void io_del(io_t *io) { - if(!io->cb) { - return; - } - - io_set(io, 0); -#ifdef HAVE_MINGW +splay_tree_t io_tree = {.compare = (splay_compare_t)io_compare}; +splay_tree_t timeout_tree = {.compare = (splay_compare_t)timeout_compare}; - if(io->fd != -1 && WSACloseEvent(io->event) == FALSE) { - abort(); - } - - event_count--; -#endif - -#ifndef HAVE_SYS_EPOLL_H - splay_unlink_node(&io_tree, &io->node); -#endif - io->cb = NULL; -} - -void timeout_add(timeout_t *timeout, timeout_cb_t cb, void *data, struct timeval *tv) { +void timeout_add(timeout_t *timeout, timeout_cb_t cb, void *data, const struct timeval *tv) { timeout->cb = cb; timeout->data = data; timeout->node.data = timeout; @@ -252,7 +85,7 @@ void timeout_add(timeout_t *timeout, timeout_cb_t cb, void *data, struct timeval timeout_set(timeout, tv); } -void timeout_set(timeout_t *timeout, struct timeval *tv) { +void timeout_set(timeout_t *timeout, const struct timeval *tv) { if(timerisset(&timeout->tv)) { splay_unlink_node(&timeout_tree, &timeout->node); } @@ -280,76 +113,7 @@ void timeout_del(timeout_t *timeout) { }; } -#ifndef HAVE_MINGW - -// From Matz's Ruby -#ifndef NSIG -# define NSIG (_SIGMAX + 1) /* For QNX */ -#endif - - -static io_t signalio; -static int pipefd[2] = {-1, -1}; -static signal_t *signal_handle[NSIG + 1] = {}; - -static void signal_handler(int signum) { - unsigned char num = signum; - write(pipefd[1], &num, 1); -} - -static void signalio_handler(void *data, int flags) { - (void)data; - (void)flags; - unsigned char signum; - - if(read(pipefd[0], &signum, 1) != 1) { - return; - } - - signal_t *sig = signal_handle[signum]; - - if(sig) { - sig->cb(sig->data); - } -} - -static void pipe_init(void) { - if(!pipe(pipefd)) { - io_add(&signalio, signalio_handler, NULL, pipefd[0], IO_READ); - } -} - -void signal_add(signal_t *sig, signal_cb_t cb, void *data, int signum) { - if(sig->cb) { - return; - } - - sig->signum = signum; - sig->cb = cb; - sig->data = data; - - if(pipefd[0] == -1) { - pipe_init(); - } - - signal(signum, signal_handler); - - signal_handle[signum] = sig; -} - -void signal_del(signal_t *sig) { - if(!sig->cb) { - return; - } - - signal(sig->signum, SIG_DFL); - - signal_handle[sig->signum] = NULL; - sig->cb = NULL; -} -#endif - -static struct timeval *timeout_execute(struct timeval *diff) { +struct timeval *timeout_execute(struct timeval *diff) { gettimeofday(&now, NULL); struct timeval *tv = NULL; @@ -371,225 +135,3 @@ static struct timeval *timeout_execute(struct timeval *diff) { return tv; } - -bool event_loop(void) { - running = true; - -#ifndef HAVE_MINGW - -#ifdef HAVE_SYS_EPOLL_H - - if(!epollset) { - epollset = event_epoll_init(); - } - -#else - fd_set readable; - fd_set writable; -#endif - - while(running) { - struct timeval diff; - struct timeval *tv = timeout_execute(&diff); -#ifndef HAVE_SYS_EPOLL_H - memcpy(&readable, &readfds, sizeof(readable)); - memcpy(&writable, &writefds, sizeof(writable)); -#endif - - -#ifdef HAVE_SYS_EPOLL_H - struct epoll_event events[EPOLL_MAX_EVENTS_PER_LOOP]; - long timeout = (tv->tv_sec * 1000) + (tv->tv_usec / 1000); - - if(timeout > INT_MAX) { - timeout = INT_MAX; - } - - int n = epoll_wait(epollset, events, EPOLL_MAX_EVENTS_PER_LOOP, (int)timeout); -#else - int maxfds = 0; - - if(io_tree.tail) { - io_t *last = io_tree.tail->data; - maxfds = last->fd + 1; - } - - int n = select(maxfds, &readable, &writable, NULL, tv); -#endif - - if(n < 0) { - if(sockwouldblock(sockerrno)) { - continue; - } else { - return false; - } - } - - if(!n) { - continue; - } - - unsigned int curgen = io_tree.generation; - - -#ifdef HAVE_SYS_EPOLL_H - - for(int i = 0; i < n; i++) { - io_t *io = events[i].data.ptr; - - if(events[i].events & EPOLLOUT && io->flags & IO_WRITE) { - io->cb(io->data, IO_WRITE); - } - - if(curgen != io_tree.generation) { - break; - } - - if(events[i].events & EPOLLIN && io->flags & IO_READ) { - io->cb(io->data, IO_READ); - } - - if(curgen != io_tree.generation) { - break; - } - } - -#else - - for splay_each(io_t, io, &io_tree) { - if(FD_ISSET(io->fd, &writable)) { - io->cb(io->data, IO_WRITE); - } else if(FD_ISSET(io->fd, &readable)) { - io->cb(io->data, IO_READ); - } else { - continue; - } - - /* - There are scenarios in which the callback will remove another io_t from the tree - (e.g. closing a double connection). Since splay_each does not support that, we - need to exit the loop if that happens. That's okay, since any remaining events will - get picked up by the next select() call. - */ - if(curgen != io_tree.generation) { - break; - } - } - -#endif - } - -#else - - while(running) { - struct timeval diff; - struct timeval *tv = timeout_execute(&diff); - DWORD timeout_ms = tv ? (DWORD)(tv->tv_sec * 1000 + tv->tv_usec / 1000 + 1) : WSA_INFINITE; - - if(!event_count) { - Sleep(timeout_ms); - continue; - } - - /* - For some reason, Microsoft decided to make the FD_WRITE event edge-triggered instead of level-triggered, - which is the opposite of what select() does. In practice, that means that if a FD_WRITE event triggers, - it will never trigger again until a send() returns EWOULDBLOCK. Since the semantics of this event loop - is that write events are level-triggered (i.e. they continue firing until the socket is full), we need - to emulate these semantics by making sure we fire each IO_WRITE that is still writeable. - - Note that technically FD_CLOSE has the same problem, but it's okay because user code does not rely on - this event being fired again if ignored. - */ - unsigned int curgen = io_tree.generation; - - for splay_each(io_t, io, &io_tree) { - if(io->flags & IO_WRITE && send(io->fd, NULL, 0, 0) == 0) { - io->cb(io->data, IO_WRITE); - - if(curgen != io_tree.generation) { - break; - } - } - } - - if(event_count > WSA_MAXIMUM_WAIT_EVENTS) { - WSASetLastError(WSA_INVALID_PARAMETER); - return(false); - } - - WSAEVENT events[WSA_MAXIMUM_WAIT_EVENTS]; - io_t *io_map[WSA_MAXIMUM_WAIT_EVENTS]; - DWORD event_index = 0; - - for splay_each(io_t, io, &io_tree) { - events[event_index] = io->event; - io_map[event_index] = io; - event_index++; - } - - /* - * If the generation number changes due to event addition - * or removal by a callback we restart the loop. - */ - curgen = io_tree.generation; - - for(DWORD event_offset = 0; event_offset < event_count;) { - DWORD result = WSAWaitForMultipleEvents(event_count - event_offset, &events[event_offset], FALSE, timeout_ms, FALSE); - - if(result == WSA_WAIT_TIMEOUT) { - break; - } - - if(result < WSA_WAIT_EVENT_0 || result >= WSA_WAIT_EVENT_0 + event_count - event_offset) { - return false; - } - - /* Look up io in the map by index. */ - event_index = result - WSA_WAIT_EVENT_0 + event_offset; - io_t *io = io_map[event_index]; - - if(io->fd == -1) { - io->cb(io->data, 0); - - if(curgen != io_tree.generation) { - break; - } - } else { - WSANETWORKEVENTS network_events; - - if(WSAEnumNetworkEvents(io->fd, io->event, &network_events) != 0) { - return(false); - } - - if(network_events.lNetworkEvents & READ_EVENTS) { - io->cb(io->data, IO_READ); - - if(curgen != io_tree.generation) { - break; - } - } - - /* - The fd might be available for write too. However, if we already fired the read callback, that - callback might have deleted the io (e.g. through terminate_connection()), so we can't fire the - write callback here. Instead, we loop back and let the writable io loop above handle it. - */ - } - - /* Continue checking the rest of the events. */ - event_offset = event_index + 1; - - /* Just poll the next time through. */ - timeout_ms = 0; - } - } - -#endif - - return true; -} - -void event_exit(void) { - running = false; -}