mirror of
https://github.com/libevent/libevent.git
synced 2025-01-31 09:12:55 +08:00
d3288293fd
The different bufferevent implementations had different behavior for their timeouts. Some of them kept re-triggering the timeouts indefinitely; some disabled the event immediately the first time a timeout triggered. Some of them made the timeouts only count when the bufferevent was actively trying to read or write; some did not. The new behavior is modeled after old socket bufferevents, since they were here first and their behavior is relatively sane. Basically, each timeout disables the bufferevent's corresponding read or write operation when it fires. Timeouts are stopped whenever we suspend writing or reading, and reset whenever we unsuspend writing or reading. Calling bufferevent_enable resets a timeout, as does changing the timeout value.
532 lines
14 KiB
C
532 lines
14 KiB
C
/*
|
|
* Copyright (c) 2009 Niels Provos and Nick Mathewson
|
|
*
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. The name of the author may not be used to endorse or promote products
|
|
* derived from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "event-config.h"
|
|
|
|
#ifdef _EVENT_HAVE_SYS_TIME_H
|
|
#include <sys/time.h>
|
|
#endif
|
|
|
|
#include <errno.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#ifdef _EVENT_HAVE_STDARG_H
|
|
#include <stdarg.h>
|
|
#endif
|
|
#ifdef _EVENT_HAVE_UNISTD_H
|
|
#include <unistd.h>
|
|
#endif
|
|
|
|
#ifdef WIN32
|
|
#include <winsock2.h>
|
|
#include <ws2tcpip.h>
|
|
#endif
|
|
|
|
#include "event2/util.h"
|
|
#include "event2/bufferevent.h"
|
|
#include "event2/buffer.h"
|
|
#include "event2/bufferevent_struct.h"
|
|
#include "event2/event.h"
|
|
#include "log-internal.h"
|
|
#include "mm-internal.h"
|
|
#include "bufferevent-internal.h"
|
|
#include "util-internal.h"
|
|
#include "iocp-internal.h"
|
|
|
|
/* prototypes */
|
|
static int be_async_enable(struct bufferevent *, short);
|
|
static int be_async_disable(struct bufferevent *, short);
|
|
static void be_async_destruct(struct bufferevent *);
|
|
static int be_async_flush(struct bufferevent *, short, enum bufferevent_flush_mode);
|
|
static int be_async_ctrl(struct bufferevent *, enum bufferevent_ctrl_op, union bufferevent_ctrl_data *);
|
|
|
|
const struct bufferevent_ops bufferevent_ops_async = {
|
|
"socket_async",
|
|
0,
|
|
be_async_enable,
|
|
be_async_disable,
|
|
be_async_destruct,
|
|
_bufferevent_generic_adj_timeouts,
|
|
be_async_flush,
|
|
be_async_ctrl,
|
|
};
|
|
|
|
struct bufferevent_async {
|
|
struct bufferevent_private bev;
|
|
struct event_overlapped connect_overlapped;
|
|
struct event_overlapped read_overlapped;
|
|
struct event_overlapped write_overlapped;
|
|
unsigned read_in_progress : 1;
|
|
unsigned write_in_progress : 1;
|
|
unsigned ok : 1;
|
|
};
|
|
|
|
static inline struct bufferevent_async *
|
|
upcast(struct bufferevent *bev)
|
|
{
|
|
struct bufferevent_async *bev_a;
|
|
if (bev->be_ops != &bufferevent_ops_async)
|
|
return NULL;
|
|
bev_a = EVUTIL_UPCAST(bev, struct bufferevent_async, bev.bev);
|
|
return bev_a;
|
|
}
|
|
|
|
static inline struct bufferevent_async *
|
|
upcast_connect(struct event_overlapped *eo)
|
|
{
|
|
struct bufferevent_async *bev_a;
|
|
bev_a = EVUTIL_UPCAST(eo, struct bufferevent_async, connect_overlapped);
|
|
EVUTIL_ASSERT(BEV_IS_ASYNC(&bev_a->bev.bev));
|
|
return bev_a;
|
|
}
|
|
|
|
static inline struct bufferevent_async *
|
|
upcast_read(struct event_overlapped *eo)
|
|
{
|
|
struct bufferevent_async *bev_a;
|
|
bev_a = EVUTIL_UPCAST(eo, struct bufferevent_async, read_overlapped);
|
|
EVUTIL_ASSERT(BEV_IS_ASYNC(&bev_a->bev.bev));
|
|
return bev_a;
|
|
}
|
|
|
|
static inline struct bufferevent_async *
|
|
upcast_write(struct event_overlapped *eo)
|
|
{
|
|
struct bufferevent_async *bev_a;
|
|
bev_a = EVUTIL_UPCAST(eo, struct bufferevent_async, write_overlapped);
|
|
EVUTIL_ASSERT(BEV_IS_ASYNC(&bev_a->bev.bev));
|
|
return bev_a;
|
|
}
|
|
|
|
static void
|
|
bev_async_consider_writing(struct bufferevent_async *b)
|
|
{
|
|
size_t at_most;
|
|
int limit;
|
|
/* Don't write if there's a write in progress, or we do not
|
|
* want to write. */
|
|
if (!b->ok || b->write_in_progress || !(b->bev.bev.enabled&EV_WRITE))
|
|
return;
|
|
/* Don't write if there's nothing to write */
|
|
if (!evbuffer_get_length(b->bev.bev.output))
|
|
return;
|
|
|
|
at_most = evbuffer_get_length(b->bev.bev.output);
|
|
|
|
/* XXXX This over-commits. */
|
|
limit = _bufferevent_get_write_max(&b->bev);
|
|
if (at_most >= limit)
|
|
at_most = limit;
|
|
|
|
if (b->bev.write_suspended)
|
|
return;
|
|
|
|
/* XXXX doesn't respect low-water mark very well. */
|
|
if (evbuffer_launch_write(b->bev.bev.output, at_most,
|
|
&b->write_overlapped)) {
|
|
EVUTIL_ASSERT(0);/* XXX act sensibly. */
|
|
} else {
|
|
b->write_in_progress = 1;
|
|
}
|
|
}
|
|
|
|
static void
|
|
bev_async_consider_reading(struct bufferevent_async *b)
|
|
{
|
|
size_t cur_size;
|
|
size_t read_high;
|
|
size_t at_most;
|
|
int limit;
|
|
/* Don't read if there is a read in progress, or we do not
|
|
* want to read. */
|
|
if (!b->ok || b->read_in_progress || !(b->bev.bev.enabled&EV_READ))
|
|
return;
|
|
|
|
/* Don't read if we're full */
|
|
cur_size = evbuffer_get_length(b->bev.bev.input);
|
|
read_high = b->bev.bev.wm_read.high;
|
|
if (read_high) {
|
|
if (cur_size >= read_high)
|
|
return;
|
|
at_most = read_high - cur_size;
|
|
} else {
|
|
at_most = 16384; /* FIXME totally magic. */
|
|
}
|
|
|
|
/* XXXX This over-commits. */
|
|
limit = _bufferevent_get_read_max(&b->bev);
|
|
if (at_most >= limit)
|
|
at_most = limit;
|
|
|
|
if (b->bev.read_suspended)
|
|
return;
|
|
|
|
if (evbuffer_launch_read(b->bev.bev.input, at_most,
|
|
&b->read_overlapped)) {
|
|
EVUTIL_ASSERT(0);
|
|
} else {
|
|
b->read_in_progress = 1;
|
|
}
|
|
}
|
|
|
|
static void
|
|
be_async_outbuf_callback(struct evbuffer *buf,
|
|
const struct evbuffer_cb_info *cbinfo,
|
|
void *arg)
|
|
{
|
|
struct bufferevent *bev = arg;
|
|
struct bufferevent_async *bev_async = upcast(bev);
|
|
|
|
/* If we added data to the outbuf and were not writing before,
|
|
* we may want to write now. */
|
|
|
|
_bufferevent_incref_and_lock(bev);
|
|
|
|
if (cbinfo->n_added)
|
|
bev_async_consider_writing(bev_async);
|
|
|
|
_bufferevent_decref_and_unlock(bev);
|
|
}
|
|
|
|
static void
|
|
be_async_inbuf_callback(struct evbuffer *buf,
|
|
const struct evbuffer_cb_info *cbinfo,
|
|
void *arg)
|
|
{
|
|
struct bufferevent *bev = arg;
|
|
struct bufferevent_async *bev_async = upcast(bev);
|
|
|
|
/* If we drained data from the inbuf and were not reading before,
|
|
* we may want to read now */
|
|
|
|
_bufferevent_incref_and_lock(bev);
|
|
|
|
if (cbinfo->n_deleted)
|
|
bev_async_consider_reading(bev_async);
|
|
|
|
_bufferevent_decref_and_unlock(bev);
|
|
}
|
|
|
|
static int
|
|
be_async_enable(struct bufferevent *buf, short what)
|
|
{
|
|
struct bufferevent_async *bev_async = upcast(buf);
|
|
|
|
if (!bev_async->ok)
|
|
return -1;
|
|
|
|
/* NOTE: This interferes with non-blocking connect */
|
|
if (event & EV_READ)
|
|
BEV_RESET_GENERIC_READ_TIMEOUT(bev);
|
|
if (event & EV_WRITE)
|
|
BEV_RESET_GENERIC_WRITE_TIMEOUT(bev);
|
|
|
|
/* If we newly enable reading or writing, and we aren't reading or
|
|
writing already, consider launching a new read or write. */
|
|
|
|
if (what & EV_READ)
|
|
bev_async_consider_reading(bev_async);
|
|
if (what & EV_WRITE)
|
|
bev_async_consider_writing(bev_async);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
be_async_disable(struct bufferevent *bev, short what)
|
|
{
|
|
/* XXXX If we disable reading or writing, we may want to consider
|
|
* canceling any in-progress read or write operation, though it might
|
|
* not work. */
|
|
|
|
if (event & EV_READ)
|
|
BEV_DEL_GENERIC_READ_TIMEOUT(bev);
|
|
if (event & EV_WRITE)
|
|
BEV_DEL_GENERIC_WRITE_TIMEOUT(bev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
be_async_destruct(struct bufferevent *bev)
|
|
{
|
|
struct bufferevent_private *bev_p = BEV_UPCAST(bev);
|
|
evutil_socket_t fd;
|
|
|
|
EVUTIL_ASSERT(!upcast(bev)->write_in_progress && !upcast(bev)->read_in_progress);
|
|
|
|
/* XXX cancel any outstanding I/O operations */
|
|
fd = _evbuffer_overlapped_get_fd(bev->input);
|
|
/* delete this in case non-blocking connect was used */
|
|
event_del(&bev->ev_write);
|
|
if (bev_p->options & BEV_OPT_CLOSE_ON_FREE)
|
|
EVUTIL_CLOSESOCKET(fd);
|
|
_bufferevent_del_generic_timeout_cbs(bev);
|
|
}
|
|
|
|
static int
|
|
be_async_flush(struct bufferevent *bev, short what,
|
|
enum bufferevent_flush_mode mode)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
connect_complete(struct event_overlapped *eo, ev_uintptr_t key,
|
|
ev_ssize_t nbytes, int ok)
|
|
{
|
|
struct bufferevent_async *bev_a = upcast_connect(eo);
|
|
struct bufferevent *bev = &bev_a->bev.bev;
|
|
|
|
_bufferevent_incref_and_lock(bev);
|
|
|
|
EVUTIL_ASSERT(bev_a->bev.connecting);
|
|
bev_a->bev.connecting = 0;
|
|
|
|
bufferevent_async_set_connected(bev);
|
|
_bufferevent_run_eventcb(bev,
|
|
ok? BEV_EVENT_CONNECTED : BEV_EVENT_ERROR);
|
|
|
|
_bufferevent_decref_and_unlock(bev);
|
|
}
|
|
|
|
static void
|
|
read_complete(struct event_overlapped *eo, ev_uintptr_t key,
|
|
ev_ssize_t nbytes, int ok)
|
|
{
|
|
struct bufferevent_async *bev_a = upcast_read(eo);
|
|
struct bufferevent *bev = &bev_a->bev.bev;
|
|
short what = BEV_EVENT_READING;
|
|
|
|
_bufferevent_incref_and_lock(bev);
|
|
EVUTIL_ASSERT(bev_a->ok && bev_a->read_in_progress);
|
|
|
|
evbuffer_commit_read(bev->input, nbytes);
|
|
bev_a->read_in_progress = 0;
|
|
|
|
if (ok && nbytes) {
|
|
BEV_RESET_GENERIC_READ_TIMEOUT(bev);
|
|
_bufferevent_decrement_read_buckets(&bev_a->bev, nbytes);
|
|
if (evbuffer_get_length(bev->input) >= bev->wm_read.low)
|
|
_bufferevent_run_readcb(bev);
|
|
bev_async_consider_reading(bev_a);
|
|
} else if (!ok) {
|
|
what |= BEV_EVENT_ERROR;
|
|
bev_a->ok = 0;
|
|
_bufferevent_run_eventcb(bev, what);
|
|
} else if (!nbytes) {
|
|
what |= BEV_EVENT_EOF;
|
|
bev_a->ok = 0;
|
|
_bufferevent_run_eventcb(bev, what);
|
|
}
|
|
|
|
_bufferevent_decref_and_unlock(bev);
|
|
}
|
|
|
|
static void
|
|
write_complete(struct event_overlapped *eo, ev_uintptr_t key,
|
|
ev_ssize_t nbytes, int ok)
|
|
{
|
|
struct bufferevent_async *bev_a = upcast_write(eo);
|
|
struct bufferevent *bev = &bev_a->bev.bev;
|
|
short what = BEV_EVENT_WRITING;
|
|
|
|
_bufferevent_incref_and_lock(bev);
|
|
EVUTIL_ASSERT(bev_a->ok && bev_a->write_in_progress);
|
|
|
|
evbuffer_commit_write(bev->output, nbytes);
|
|
bev_a->write_in_progress = 0;
|
|
|
|
if (ok && nbytes) {
|
|
BEV_RESET_GENERIC_WRITE_TIMEOUT(bev);
|
|
_bufferevent_decrement_write_buckets(&bev_a->bev, nbytes);
|
|
if (evbuffer_get_length(bev->output) <= bev->wm_write.low)
|
|
_bufferevent_run_writecb(bev);
|
|
bev_async_consider_writing(bev_a);
|
|
} else if (!ok) {
|
|
what |= BEV_EVENT_ERROR;
|
|
bev_a->ok = 0;
|
|
_bufferevent_run_eventcb(bev, what);
|
|
} else if (!nbytes) {
|
|
what |= BEV_EVENT_EOF;
|
|
bev_a->ok = 0;
|
|
_bufferevent_run_eventcb(bev, what);
|
|
}
|
|
|
|
_bufferevent_decref_and_unlock(bev);
|
|
}
|
|
|
|
struct bufferevent *
|
|
bufferevent_async_new(struct event_base *base,
|
|
evutil_socket_t fd, int options)
|
|
{
|
|
struct bufferevent_async *bev_a;
|
|
struct bufferevent *bev;
|
|
struct event_iocp_port *iocp;
|
|
|
|
options |= BEV_OPT_THREADSAFE;
|
|
|
|
if (!(iocp = event_base_get_iocp(base)))
|
|
return NULL;
|
|
|
|
if (fd >= 0 && event_iocp_port_associate(iocp, fd, 1)<0) {
|
|
int err = GetLastError();
|
|
/* We may have alrady associated this fd with a port.
|
|
* Let's hope it's this port, and that the error code
|
|
* for doing this neer changes. */
|
|
if (err != ERROR_INVALID_PARAMETER)
|
|
return NULL;
|
|
}
|
|
|
|
if (!(bev_a = mm_calloc(1, sizeof(struct bufferevent_async))))
|
|
return NULL;
|
|
|
|
bev = &bev_a->bev.bev;
|
|
if (!(bev->input = evbuffer_overlapped_new(fd))) {
|
|
mm_free(bev_a);
|
|
return NULL;
|
|
}
|
|
if (!(bev->output = evbuffer_overlapped_new(fd))) {
|
|
evbuffer_free(bev->input);
|
|
mm_free(bev_a);
|
|
return NULL;
|
|
}
|
|
|
|
if (bufferevent_init_common(&bev_a->bev, base, &bufferevent_ops_async,
|
|
options)<0)
|
|
goto err;
|
|
|
|
evbuffer_add_cb(bev->input, be_async_inbuf_callback, bev);
|
|
evbuffer_add_cb(bev->output, be_async_outbuf_callback, bev);
|
|
evbuffer_defer_callbacks(bev->input, base);
|
|
evbuffer_defer_callbacks(bev->output, base);
|
|
|
|
event_overlapped_init(&bev_a->connect_overlapped, connect_complete);
|
|
event_overlapped_init(&bev_a->read_overlapped, read_complete);
|
|
event_overlapped_init(&bev_a->write_overlapped, write_complete);
|
|
|
|
bev_a->ok = fd >= 0;
|
|
|
|
return bev;
|
|
err:
|
|
bufferevent_free(&bev_a->bev.bev);
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
bufferevent_async_set_connected(struct bufferevent *bev)
|
|
{
|
|
struct bufferevent_async *bev_async = upcast(bev);
|
|
bev_async->ok = 1;
|
|
_bufferevent_init_generic_timeout_cbs(bev);
|
|
/* Now's a good time to consider reading/writing */
|
|
be_async_enable(bev, bev->enabled);
|
|
}
|
|
|
|
int
|
|
bufferevent_async_can_connect(struct bufferevent *bev)
|
|
{
|
|
const struct win32_extension_fns *ext =
|
|
event_get_win32_extension_fns();
|
|
|
|
if (BEV_IS_ASYNC(bev) &&
|
|
event_base_get_iocp(bev->ev_base) &&
|
|
ext && ext->ConnectEx)
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
bufferevent_async_connect(struct bufferevent *bev, evutil_socket_t fd,
|
|
const struct sockaddr *sa, int socklen)
|
|
{
|
|
BOOL rc;
|
|
struct bufferevent_async *bev_async = upcast(bev);
|
|
struct sockaddr_storage ss;
|
|
const struct win32_extension_fns *ext =
|
|
event_get_win32_extension_fns();
|
|
|
|
EVUTIL_ASSERT(ext && ext->ConnectEx && fd >= 0 && sa != NULL);
|
|
|
|
/* ConnectEx() requires that the socket be bound to an address
|
|
* with bind() before using, otherwise it will fail. We attempt
|
|
* to issue a bind() here, taking into account that the error
|
|
* code is set to WSAEINVAL when the socket is already bound. */
|
|
memset(&ss, 0, sizeof(ss));
|
|
if (sa->sa_family == AF_INET) {
|
|
struct sockaddr_in *sin = (struct sockaddr_in *)&ss;
|
|
sin->sin_family = AF_INET;
|
|
sin->sin_addr.s_addr = INADDR_ANY;
|
|
} else if (sa->sa_family == AF_INET6) {
|
|
struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&ss;
|
|
sin6->sin6_family = AF_INET6;
|
|
sin6->sin6_addr = in6addr_any;
|
|
} else {
|
|
/* Well, the user will have to bind() */
|
|
return -1;
|
|
}
|
|
if (bind(fd, (struct sockaddr *)&ss, sizeof(ss)) < 0 &&
|
|
WSAGetLastError() != WSAEINVAL)
|
|
return -1;
|
|
|
|
rc = ext->ConnectEx(fd, sa, socklen, NULL, 0, NULL,
|
|
&bev_async->connect_overlapped.overlapped);
|
|
if (rc || WSAGetLastError() == ERROR_IO_PENDING)
|
|
return 0;
|
|
|
|
return -1;
|
|
}
|
|
|
|
static int
|
|
be_async_ctrl(struct bufferevent *bev, enum bufferevent_ctrl_op op,
|
|
union bufferevent_ctrl_data *data)
|
|
{
|
|
switch (op) {
|
|
case BEV_CTRL_GET_FD:
|
|
data->fd = _evbuffer_overlapped_get_fd(bev->input);
|
|
return 0;
|
|
case BEV_CTRL_SET_FD: {
|
|
struct event_iocp_port *iocp;
|
|
|
|
if (data->fd == _evbuffer_overlapped_get_fd(bev->input))
|
|
return 0;
|
|
if (!(iocp = event_base_get_iocp(bev->ev_base)))
|
|
return -1;
|
|
if (event_iocp_port_associate(iocp, data->fd, 1) < 0)
|
|
return -1;
|
|
_evbuffer_overlapped_set_fd(bev->input, data->fd);
|
|
_evbuffer_overlapped_set_fd(bev->output, data->fd);
|
|
return 0;
|
|
}
|
|
case BEV_CTRL_GET_UNDERLYING:
|
|
default:
|
|
return -1;
|
|
}
|
|
}
|